COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

102
COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL MAMMOGRAPHY Jo˜ ao Maria Patacho de Matos Rocha e Melo Thesis to obtain the Master of Science Degree in Electrical and Computer Engineering Supervisors: Prof. Maria Margarida Campos da Silveira Prof. Nuno Miguel de Pinto Lobo e Matela Examination Committee Chairperson: Prof. Jo ˜ ao Fernando Cardoso Silva Sequeira Supervisor: Prof. Maria Margarida Campos da Silveira Members of the Committee: Prof. Jorge dos Santos Salvador Marques May 2017

Transcript of COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Page 1: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITALMAMMOGRAPHY

Joao Maria Patacho de Matos Rocha e Melo

Thesis to obtain the Master of Science Degree in

Electrical and Computer Engineering

Supervisors: Prof. Maria Margarida Campos da SilveiraProf. Nuno Miguel de Pinto Lobo e Matela

Examination Committee

Chairperson: Prof. Joao Fernando Cardoso Silva SequeiraSupervisor: Prof. Maria Margarida Campos da Silveira

Members of the Committee: Prof. Jorge dos Santos Salvador Marques

May 2017

Page 2: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …
Page 3: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

To my parents.

For despite all the struggles a family can go through, my education was never at stake.

Page 4: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Acknowledgments

I would like to thank Professor Margarida Silveira, from IST, supervisor of this thesis. Her input

and recommendations were essential throughout this work. I would like to particularly thank her for her

patience and understanding for, due to professional reasons, I had many time constrains during this

work’s period. Professor Margarida was nothing but understanding and supportive of this fact. I also

want to thank Professor Nuno Matela, from IBEB, co-supervisor of this work. Professor Nuno’s input on

the medical subjects were essential for the completion of this project. I would like to thank him for the

several trips to the Hospital in order to collect data and experts’ opinions.

A special thank you to Professor Jose Roquette and Dr Isabel Estudante, from Luz-Saude, and Dr

Tiago Saldanha for their availability in providing expert medical opinion regarding several topics as well

as for their recommendation on the user side of this work’s final product.

I would like to thank Ana Colaco, Bsc. Ana volunteered to do an extensive research on the medical

aspects of breast cancer and shared such research with me and several other students.

I cannot thank my entire family enough, in particular my parents and my girlfriend, for their under-

standing on the work and time invested in this thesis. Many sacrifices are done by researchers, when

doing a project such as this one. Nevertheless, many are also done by the people around them and

such should be acknowledged. For this, I thank them. Also, a great thank to Ederzito A.M. Lopes. His

inspiring attitude towards his profession and his accomplishments were crucial for the conclusion of this

work.

Also, to my friends, particularly my colleagues at IST: Bernardo Almeida, Bernardo Marques, Hugo

da Silva, Joao Franco, Joao Oliveira, Jose Teixeira, Manuel Avila de Melo, Manuel Ribeiro, Manuel

Beja da Costa, Miguel Monteiro and Nuno Sousa. Their input, both technical and emotional and their

recommendations and advices were fundamental for the conclusion of both this thesis and my studies.

I would like to thank my company, BNP Paribas, my team and my co-workers, for their support and

for letting me dedicate time to this project, when it was necessary.

Last, but not least, I would like to thank all my teachers, Professors and everyone involved in my

education whom, without their support, my journey towards here would have been different and harder.

To all of you, thank you.

Page 5: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Abstract

Currently, early detection of breast cancer is one of the most important factors in the treatment of breast

cancer. Computer Aided Diagnosis (CAD) systems provide aid to doctors in diagnosing breast images

such as mammography and Digital Breast Tomosynthesis (DBT), in order to increase the diagnosis’

accuracy. This work describes a CAD system to locate lesions (masses) in 2D mammographies, based

on image processing and machine learning techniques. This thesis proposes a two stage combination of

classifiers to label mammograms. The first stage separates masses from healthy tissue while the second

distinguishes between malignant or benign masses. Each base classifier is trained with a different type

of feature, either Morphological, Intensity or Texture based. Two base classifiers, Adaboost and Decision

Trees, were tested. Results obtained on real mammograms from an international database showed an

improvement in the accuracy of the CAD system, from 82% to 85%, when compared to the use of a

single classifier. In this thesis, an hypothesis of a 3 model ensemble classifier is implemented to label

a database of mammograms into healthy tissue, malignant mass or benign mass. In this classification

stage, two classifiers (AdaBoost and Decision Tree) were tested. This hypothesis proved to increase the

accuracy of the CAD system, when comparing to a single model classifier, from 82% to 85%.

Keywords

Mammography; Breast Cancer; Computer Aided Diagnosis.

iii

Page 6: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …
Page 7: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Resumo

Nos dias de hoje, a detecao em estado inicial do cancro da mama e um dos aspectos mais decisivos

para minimizar os impactos e consequencias deste tipo de cancro. Sistemas de Diagnostico Auxiliado

por Computador providenciam suporte aos medicos, aquando o diagnostico via tecnicas de imagiologia

como mamografia e tomossıntese, para aumentar a precisao do diagnostico. Este trabalho descreve

um sistema de diagnostico auxiliado por computador para localizar lesoes (massas), em mamografias

2D, baseado em tecnicas de processamento de imagem e aprendizagem automatica. Esta dividido

em seis capıtulos: uma introducao as tecnicas de imagiologia, do ponto de vista medico e tecnico, um

resumo do estado da arte dos sistemas de diagnostico auxiliado por computador para mamografias

e tomossınteses, uma explicacao extensiva dos metodos usados para a construcao do sistema, os

resultados do mesmo e uma conclusao e futuras melhorias para a continuacao deste trabalho. Nesta

dissertacao, uma hipotese baseada em fundir 3 modelos num unico classificador e implementada, para

diagnosticar mamografias presentes numa basa de dados em: tecido saudavel, massa maligna ou

massa benigna. Nesta etapa de classificacao, dois classificadores (AdaBoost e arvores de decisao)

e tres tecnicas de fusao sao testados. Comprovou-se que esta hipotese melhora os resultados do

diagnostico auxiliado por computador, quando comparado com a implementacao de apenas um modelo,

de 82% para 85%.

Palavras Chave

Mamografia; Cancro da Mama; Diagnostico Auxiliado por Computador;

v

Page 8: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …
Page 9: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Contents

1 Introduction 1

1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.2 Breast and Breast Tumours . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.2.1 The breast and its anatomy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.2.2 Breast Tumours . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6

1.2.3 BI-RADS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6

1.3 Breast Imaging . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

1.3.1 Mammography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

1.3.2 Tomosynthesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9

1.4 Computer Aided Diagnosis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

2 State-of-the-Art 13

2.1 Pre-processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

2.2 Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17

2.3 Feature Extraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

2.4 Learning and Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

2.5 Summary of State-of-the-Art . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

3 Pre-Processing and Segmentation 27

3.1 Pre-Processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

3.2 Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

4 Feature Extraction and Classification 41

4.1 Feature Extraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43

4.1.1 Morphological Features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43

4.1.2 Grey Level Features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46

4.1.3 Texture Features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

4.2 Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

4.2.1 Decision Tree . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51

4.2.2 AdaBoost . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55

vii

Page 10: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

4.2.3 Fusion Techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60

5 Experimental Results 63

5.1 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

5.2 Pre-Processing and Segmentation Results . . . . . . . . . . . . . . . . . . . . . . . . . . 66

5.3 Classification Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69

5.3.1 Single Model Decision Tree Experiment . . . . . . . . . . . . . . . . . . . . . . . . 69

5.3.2 Decision Tree Ensemble Experiment . . . . . . . . . . . . . . . . . . . . . . . . . . 70

5.3.3 AdaBoost Ensemble Experiment . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72

5.4 Comparison of Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73

6 Conclusions and Future Work 77

viii

Page 11: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

List of Figures

1.1 Relative survivals of female patients diagnosed with breast cancer up to one year earlier,

in Great Britain, brokedown by age category. . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.2 The anatomy of the breast. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.3 Mammography performed on the left breast of a patient with a class 1 density breast. . . 8

1.4 Example of the 2 different perspectives taken on a mammography. . . . . . . . . . . . . . 8

1.5 Drawing of a mammograph examining the right breast of a patient. . . . . . . . . . . . . . 8

1.6 Angular capability of Siemens Model. This particular model creates up to 25 images. . . . 10

1.7 Example of a tomosynthesis exam where the relationship between the different slices can

be seen. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10

1.8 CAD output displayed where the previously marked lesion can be seen in green and the

CAD suggestion in red. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

2.1 Example of a bilateral filter in which the centre in question is in a high intensity area but

close to the border. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

2.2 Example of a COIN filter being computed for a circle. . . . . . . . . . . . . . . . . . . . . . 18

2.3 Representation of the computation of the CRI. . . . . . . . . . . . . . . . . . . . . . . . . 20

2.4 Representation of a labelled dataset and a new pattern, waiting to be classified. . . . . . 22

3.1 Scheme of the 4 steps approach applied in this thesis . . . . . . . . . . . . . . . . . . . . 29

3.2 Example of a a) digitalized mammography from Digital Database for Screening Mammog-

raphy (DDSM) and b) its histogram. As explained before, it can be seen that even within

the breast, darker areas may exist . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

3.3 Example of a) mammography and b) its histogram, where a Gaussian filter with a standard

deviation of 50 is applied. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

ix

Page 12: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

3.4 Binary image obtained from dividing the mammography presented in 3.2(a) into back-

ground and foreground. Notice on the top right corner of a) that other elements were

wrongly identified as foreground. Hence the need to withdraw only the larger object as

seen in b). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

3.5 Resultant image of separating breast (foreground) from background using Otsu’s method.

The green line represents the border of the mask computed from this breast. This process

is important for many of the methods following this one will be applied only to pixels within

the mask. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

3.6 2D spatial filter represented in a) 3D and b) from the x-y plane. In this figure, it is possible

to see that the further way a pixel is of the pixel of interest (center), the less its weight is

going to be. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

3.7 2D intensity filter representation in a) 3D and b) from the x-y plane. The pixels with the

highest weight value represented in this filter are the ones with the intensity value closest

to the value of the pixel of interest. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

3.8 2D bilateral filter representation in a) 3D and b) in x-y plane. Combination of filters repre-

sented in figures 3.6 and 3.7, where it can be seen that pixels closer to the pixel of interest

in terms of spatial distance and intensity level value will have higher weights. . . . . . . . 35

3.9 Comparison between image from the database used in this thesis (to the left) and micro-

scopic fluorescent image (to the right) segmented using SBF . . . . . . . . . . . . . . . . 36

3.10 Gradient representation in a a) black and white image and b) its quiver representation

from Matlab . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

3.11 Graphical representation of the gradient of a pixel computed using Intermediate Differ-

ences gradient . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37

3.12aGradient of an image. Each arrow represents the gradient vector in each pixel of the

surrounding area of the pixel of interest. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

3.12bLine segments uniting each pixel with the pixel of interest. . . . . . . . . . . . . . . . . . . 38

3.13 Fusion of the gradient of each pixel and the line segment uniting each pixel with the pixel of

interest. Graphical representation of the angle used to compute the degree of convergence. 38

3.14 Illustration of Convergence Index (COIN) filter being computed for an approximately cir-

cular lesion. This figure exemplifies a case where the centre is correctly found, but the

border is not. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

3.15 Illustration of an iris filter being computed for the same approximately circular lesion as

in figure 3.14. This figure exemplifies a case where both the centre and the border are

found. In this illustration, only 8 orientation points are used, hence, the border is not

entirely found, although, if N →∞ , it would. . . . . . . . . . . . . . . . . . . . . . . . . . 40

x

Page 13: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

3.16 Illustration of an adaptive ring filter being computed for the same approximately circular

lesion as in figure 3.14 and 3.15. In this illustration, the Rmin and Rmax are used to limit

the area in which the border can be found. . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

4.1 Mathematical morphology operation a) ”dilatation” and b) ”erosion” performed between

image A, a square, and the structuring element B, a disk. . . . . . . . . . . . . . . . . . . 44

4.2 Example of a round region where mathematical morphology opening was applied. At it

can be visualized, the image does not differ from its original one. . . . . . . . . . . . . . . 45

4.3 On the contrary of the region presented in figure 4.2, when opening is applied to an

approximately round region, some changes can be observed already. . . . . . . . . . . . 45

4.4 Opposing the two cases earlier, figures 4.2 and 4.3, when opening is applied to a spicu-

lated region, a significant difference can be seen. . . . . . . . . . . . . . . . . . . . . . . . 46

4.5 Examples of histograms of Region of Interest (ROI)s with the same mean intensity value

but with different levels of standard deviation. On the left, an histogram with a small value

of standard deviation. On the right, an histogram with a standard deviation value 5 times

higher than the one on the left. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

4.6a Example of the angle’s histogram of an image. On the left, the histogram of an image

with a random texture (presented on the right). By this histogram, it is possible to see that

there are gradients pointing to all directions in the image on the left. . . . . . . . . . . . . 49

4.6b Illustration of how an analysis on the histogram of the gradients’ angle of an image can be

descriptive of such image. In this case, on the right an image with stripes and on the left,

its histogram. It can be easily seen that, comparing to figure 4.6a, this histogram outputs

more information regarding the organization of its image. . . . . . . . . . . . . . . . . . . 49

4.7 Representation of the classification flow used in this work. The features of the ROI are

used as input for 3 classifiers, whose outputs are combined and in which the ROI is

classified as a Mass or as Healthy Tissue. In case of a mass, the process moves along

and the features are used as inputs to another 3 classifiers, that, after its outputs are

combined, define if the ROI is malignant lesion or a benign one. . . . . . . . . . . . . . . . 51

4.8 Example of a DT with three nodes and four leafs. . . . . . . . . . . . . . . . . . . . . . . . 52

4.9 Gini function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52

4.10 Ilustration of the Single Model Decision Tree used in this work . . . . . . . . . . . . . . . . 53

4.11 Schema of the classifier used in the Decision Tree Ensemble experiment. . . . . . . . . . 53

4.12 Ilustration of the AdaBoost Classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55

4.13 Illustration of the first learner being computed. In this example, all the patterns have the

same weight (represented by the patterns size) . . . . . . . . . . . . . . . . . . . . . . . . 56

xi

Page 14: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

4.14 Illustration of the second learner of an AdaBoost. In this figure, the patterns previously

miss-classified in figure 4.13 have a larger classification weight, as represented by their

size. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57

4.15 Illustration of the third learner of the AdaBoost. As it can be seen, the patterns miss

classified both in figure 4.13 or 4.14 have larger weights than the others. . . . . . . . . . . 57

4.16 Schema of the classifier used in the AdaBoost Ensemble experiment. . . . . . . . . . . . 58

5.1 Histogram of the 3 classes of exams contained in DDSM database. . . . . . . . . . . . . . 65

5.2 Effect of filtering on the segmentation of a breast image with a malignant lesion a) without

the bilateral filter b) with the bilateral filter. In this figure, it can be seen the output of the

segmentation in green and the ground truth in blue. . . . . . . . . . . . . . . . . . . . . . . 67

5.3 Effect of filtering on the segmentation of a breast image with a malignant lesion a) without

the bilateral filter b) with the bilateral filter. In this case, the filter had no effect on the

segmentation output. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68

5.4 Effect of filtering on the segmentation of a breast image with a malignant lesion a) without

the bilateral filter b) with the bilateral filter. In this case, the segmentation algorithm does

not locate the correct ROI in the filtered image. . . . . . . . . . . . . . . . . . . . . . . . . 68

5.5 DT a) not pruned, b) pruned, with three possible outputs: 0,1 and 2 or healthy tissue,

malignant mass and benign mass. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69

5.6 Relative weights of the learners of the AdaBoost classifier a) in total b) by model. . . . . . 73

xii

Page 15: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

List of Tables

1.1 Percentage value of the 3 most common types of cancer. The values represent the

amount of cases of a certain type of cancer in a universe of cancer diagnosed patients. . 3

2.1 Table presenting chronologically several important contributions in the field of CAD of

breast cancer in mammography and DBT in the past recent years . . . . . . . . . . . . . 25

5.1 The female population at MGH and WFUSM by race . . . . . . . . . . . . . . . . . . . . . 66

5.2 Confusion matrix of the decision tree presented in the Single Model Decision Tree exper-

iment. This model obtained a classification rate of 82%. . . . . . . . . . . . . . . . . . . . 70

5.3 Table presenting the per class accuracy results of the Single Model Decision Tree experi-

ment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70

5.4 Confusion matrix of the classifier presented in the Decision Tree Ensemble Experiment.

The final classifier obtained a classification rate of 82%. . . . . . . . . . . . . . . . . . . . 71

5.5 Table presenting the accuracies per class and per model obtained from the results of the

Decision Tree Ensemble Experiment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71

5.6 Confusion matrix of the morphological, intensity and texture features model presented in

the Decision Tree Ensemble Experiment. Each model obtained a classification rate of

80%, 73% and 71%, respectively. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71

5.7 Confusion matrix of the AdaBoost Ensemble Experiment classifier. This model obtained

a classification rate of 85%. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72

5.8 Table presenting the accuracies per class and per model obtained from the results of the

AdaBoost Ensemble Experiment classifier. . . . . . . . . . . . . . . . . . . . . . . . . . . 73

5.9 Table for comparison of general accuracies and per class accuracy of the three experiments 74

5.10 Confusion matrix for all three experiments, considering only two masses. . . . . . . . . . 75

xiii

Page 16: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

xiv

Page 17: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Acronyms

ACC Accuracy

AT Adaptive Thresholding

AUC Area Under the Curve

BI-RADS Breast Imaging Reporting and Data System

CAD Computer Aided Diagnosis

CC Craniocaudal

CRI Candidate Ratio Index

COIN Convergence Index

DBT Digital Breast Tomosynthesis

DDSM Digital Database for Screening Mammography

DT Decision Tree

EM Expectation Maximization

FCM Fuzzy C-Means

FFDM Full-Field Digital Mammography

FP False Positive

GFA Gradient Field Analysis

IBEB Instituto de Biofısica e Engenharia Biomedica

IEEE Institute of Electrical and Electronics Engineers

ICIP International Conference on Image Processing

xv

Page 18: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

IST Instituto Superior Tecnico

IJBBE International Journal of Biology and Biomedical Engineering

HOG Histogram of Oriented Gradients

LCF Local Convergence Filters

LDA Linear Discriminant Analysis

MGH Massachusetts General Hospital

MRI Magnetic Resonance Imaging

MLO Mediolateral Oblique

MLP Multi-layer Perceptron

MP Medical Physics

PV Projection View

RBF Radial Basis Function

RGI Radial Gradient Index

RLS Run Length Statistics

ROI Region of Interest

SBF Sliding Band Filter

SC Surface Change

SGLD Spatial Grey Level Dependence

SPIE Society of Photo-Optical Instrumentation Engineers

SSD Sum of Squared Differences

SVM Support Vector Machine

TP True Positive

VOI Volume of Interest

WFUSM Wake Forest University School of Medicine

xvi

Page 19: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

1Introduction

Contents

1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.2 Breast and Breast Tumours . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.3 Breast Imaging . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

1.4 Computer Aided Diagnosis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

1

Page 20: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

2

Page 21: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

1.1 Motivation

According to the World Health Organization, cancer incidence worldwide increased from 12.7 million

cases in 2008 to 14.1 million in 2012. If this increase continues, this will bring the number of cancer

cases close to 25 million over the next two decades [1]. According to the same source, in the year of

2012, 8.2 million people died of cancer, which represents a cancer mortality rate of 64.5%. Despite the

fact that lung cancer is the most common type of cancer [2] and the one with the lowest survival rate [3],

breast cancer represents 25% of all cancer cases in the female population, being the most common

cancer for women, as shown in table 1.1. 1

Table 1.1: Percentage value of the 3 most common types of cancer. The values represent the amount of cases ofa certain type of cancer in a universe of cancer diagnosed patients.

GenderBoth (%) Male (%) Female (%)

Typesof

Cancer

Breast 11.9 - 25.2Lung 13.0 16.7 8.8

Colorectum 9.7 10.0 9.2

In 2012, Breast Cancer’s survival rate was 87% for people who were diagnosed 5 years earlier

(or more) in the United States [4], and 78% for people diagnosed 10 years earlier (or more) in Great

Britain [3]. One of the most peculiar aspects of this type of cancer is that, if caught in its early stages,

the survival rate is very high, but it drops at a steep rate when diagnosed in later stages, lowering to as

much as half the value when the patient is of advanced age. Figure 1.1 exemplifies this.

Figure 1.1: Relative survivals of female patients diagnosed with breast cancer up to one year earlier, in GreatBritain, brokedown by age category. Source: Public Health England [5]

1Male population not displayed for breast cancer, for not statistically significant

3

Page 22: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

This data points towards an important conclusion. Quicker and more efficient diagnosis methods,

for breast cancer prevention and early detection, are very important and should be increasingly used,

given the amount of breast cancer cases. Currently, the most common breast imaging technique used

for breast cancer diagnosis is the mammography, a 2D X-ray technique used to visualize lesions in

the patient’s breast [6]. Being an imaging technique, mammography implies a doctor to spend time

evaluating the exam itself in order to obtain a diagnosis. This means more expenses for the hospital

(doctor’s allocated time to this task) and that fatigue and lack of time for evaluations are important

factors in the diagnosis sensitivity.

Due to all these reasons, the need for Computer Aided Diagnosis (CAD) systems to complement

mammography is expected to rise. A CAD system uses image processing techniques and machine

learning algorithms to perform an analysis on the exam and helps the doctor with its diagnosis. Other

CAD systems have already been implemented in several imaging techniques, increasing 6% to 17% the

number of correct diagnoses depending on the type of cancers.

4

Page 23: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

1.2 Breast and Breast Tumours

1.2.1 The breast and its anatomy

The breast is composed by several different regions, types of tissues and muscles, as seen in figure

1.2. On one hand, the adipose tissue is composed by muscles and fat, and it represents the less

dense sections of the breast. On the other hand, glandular tissue is denser and it is where glands and

lactiferous ducts2 can be found.

Figure 1.2: The anatomy of the breast. Source: Interventional Ultrasound of the Breast [7]

The measurement of the ratio between glandular and adipose tissue is called density. A patient’s

breast can be of density 1 to 4, with the following specifics:

• Density 1: Significantly larger presence of adipose tissue. A breast with these characteristics is

little dense and fat.

• Density 2: Equally distributed amounts of glandular and adipose tissue, in an organized and

distinct manner. In this type of breast, there is more glandular tissue in comparison with density 1

breast, but adipose and glandular tissue do not mix nor overlap.

• Density 3: Equally distributed amounts of glandular and adipose tissue, such as in density 2

breasts, but in a de-organized manner. Several regions of each type of tissue with no apparent

order or criteria are present in a breast of density 3.

• Density 4: Larger presence of glandular tissue and lactiferus ducts.

2Channel that transports breast milk from its origin (glandular cells) to the nipple

5

Page 24: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

1.2.2 Breast Tumours

A tumour is a conglomerate of cells that is out of the ordinary. Contrary to common belief, a tumour

does not necessarily pose health issues, nor troubles. There is a variety of tumours which have different

behaviours, shapes, locations, etc, and each of them is treated and analysed in a different way. A tumour

(also called lesion or mass) is of one of two types:

• Benign lesion: this lesion is encapsulated. It means it does not spread nor grows out of control.

• Malignant lesion: commonly known as cancer, these types of lesions grow uncontrolled and

disproportional to what the human body expects. These cells can invade surrounding tissue and/or

metastatize to other tissues in other parts of the body.

This thesis focus in tumours located in the breast area. As with other tumours, they can be either

malignant or benign and are more common among the female population than other tumours due to the

anatomy of the breast. The lactiferus ducts are auspicious environments for tumours to develop due to

the large number of glandes and to the presence of breast milk which contains calcium and can solidify

into lesions as well.

Lesions in the breast can be divided into two major categories, where both can be benign or malig-

nant:

• Masses: structures with an oval and round shape. Can be of various sizes.

• Microcalcifications: conglomerates of calcium originated from breast milk. These lesions are

significantly smaller than masses and can form clusters which can become malignant.

1.2.3 BI-RADS

The Breast Imaging Reporting and Data System (BI-RADS) is a classification system of the appar-

ent malignancy of the lesion of a patient. After reviewing a radiology exam (mammography, Magnetic

Resonance Imaging (MRI) or breast ultrasound), the radiologist determines, in a scale from 0 to 6, how

malignant (or not) the findings in the exams are. This classification system is divided into the following

items:

• BI-RADS 0: This level of BI-RADS indicates that the study is not complete. More information (or

more exams) are still required or undergoing.

• BI-RADS 1: This indicates that no lesion was found by visual inspection of the exam. The patient

should continue the routine screenings, but there is no factor that indicates any other action should

take place.

• BI-RADS 2: When this level is discriminated, it indicates that the doctor found an irregularity but

with no reason for further precautions. This means the patient should continue routine screenings.

6

Page 25: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

This is important because it tells the next doctor that sees the next exam, that something was

previously found. This way the doctor can check if it evolved to something worse, if it stayed the

same or if it vanished.

• BI-RADS 3: This level of BI-RADS indicates that something slightly suspicious was found in the

patient’s exam. No immediate action is required, but the patient is asked to repeat the exam in a 6

month period for further confirmation.

• BI-RADS 4: This indicates that the findings in the patient’s mammogram are suspicious of being

malignant. When this level is attributed, the patient will undergo a biopsy 3 exam for confirmation

of the malignancy of the lesion.

• BI-RADS 5: When this level is discriminated, it indicates that, in the expert’s opinion, the findings

appear malignant. This is the maximum level of BI-RADS that can be assigned to a patient without

undergoing a biopsy (only by visual inspection). Similar to level 4, the next step is to perform a

biopsy to confirm (or not) the malignancy of the lesion.

• BI-RADS 6: The highest level of BI-RADS that can be assigned to a patient is level 6. This means

the patient was diagnosed with breast cancer and should discussed (or already did) a treatment

plan with an expert.

1.3 Breast Imaging

1.3.1 Mammography

Currently, the most common and effective form of breast cancer diagnosis is the digital mammogra-

phy [8]. This exam consists in X-ray images of the breast that can be taken from two standard directions

(Craniocaudal (CC) view and Mediolateral Oblique (MLO) view) and many supplementary views [9]. As

seen in figure 1.4, a CC view mammography will obtain an image overlap from top to bottom of the

breast, compared to the MLO which will in turn obtain a 45º side-view image of the breast. These

processes are done for two reasons:

• Some sections of the breast can only be seen from one of the projections

• It may be important for the doctor to have a different perspective of a certain lesion, which can only

be done through a different view (or DBT).

As mentioned in section 1.1, breast cancer is a common issue among the female population. For

this reason, it is generally advised for women to perform annual routine mammographies from the age

of 40. Nevertheless, if a patient of any gender feels a lump in the breast region and/or has genetic

3A biopsy is an exam where a needle is used to collect a sample of the suspicious tissue. This is the highest form of confirmationof breast cancer available, excluding actual surgery.

7

Page 26: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

tendency to have breast cancer, he or she should get advice from a specialist on whether a screening

mammography should take place.

Figure 1.3: Mammography performed on the left breastof a patient with a class 1 density breast.Source: Canada Health [10]

Figure 1.4: Example of the 2 different perspectivestaken on a mammography. CC in red andMLO in blue. Source: Statistical modelsof mammographic texture and appearance[11]

Figure 1.5: Drawing of a mammograph examining the right breast of a patient. Source: Canadian Cancer Society[12]

When a mammography is to be performed, the patient’s breast must be placed in a device called

mammograph (figure 1.5). Here, the breast is compressed between the detector and the compression

plates and beamed with radiation, which will later be constructed into an image. This process is both

painful and brings health risks due to the patients’ exposition to malignant radiation levels.

The revolution of the digital era has also affected the way mammographies are performed. Previously,

silver foils were beamed with x-ray radiation, a process which would darken the foil accordingly to the

level of radiation that would reached it. This means that areas that were exposed to a lot of radiation

would turn almost black and areas with less exposition would remain silver coloured. This detects

masses and microcalcifications, for these lesions are much denser than normal breast tissue, hence,

8

Page 27: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

less radiation would go through and reach the silver foil. The final result would be a generally dark

image (where no lesion or glands are present), lighter sections where skin and glandular tissue can be

encountered and a very light region if a mass or microcalcifications existed.

Later, digital mammography was introduced, where the silver foil was replaced by an electronic circuit

with radiation micro-detectors that represent a pixel in the digital image. For the patient and for the

radiologist, the process is the same: radiation is beamed over the detector and through the breast.

This process results in different levels of radiation reaching each micro-detector. In the final result,

a micro-detector that was beamed with a high level of radiation outputs a dark pixel and vice-versa.

This convention was chosen in order to keep consistency and similarities to mammographies performed

before the digital era.

Finally, after the exam is performed by a radiology technician, it is observed by a radiologist. This

specialist will give it a BI-RADS classification (as explained in section 1.2.3), a small description and

indicates what should happen next:

• Nothing

• Keep routine mammographies

• Perform another mammography

• Undergo a biopsy exam for tissue evaluation

If the result of the biopsy is positive, the patient is diagnosed with breast cancer and he or she are

redirected to a specialist in oncology (speciality of medicine that focus on cancer treatment).

So far, mammographies’ sensitivity is around 88 to 93.1%, in general, being these values significantly

lower for denser breasts [13] [14], which indicates that even when a doctor observes a mammography

image, he or she may not detect with 100% certainty if there is a lesion or not. This thesis focus on a

computer aided diagnosis technique to improve this situation.

1.3.2 Tomosynthesis

A different breast imaging technique was introduced in the most recent years to resolve a common

issue with the mammography exam: tissue overlap. Tissue overlap happens when several layers of

denser tissues are over each other, creating the illusion of a very dense volume instead of a summation

of slightly denser tissue segments. This different technique, Digital Breast Tomosynthesis (DBT) has

been introduced due to its capability to reconstruct the breast in a 3D environment. This way, it is

possible to distinguish tissue overlap from lesions, hence improving a doctor’s ability to classify what is

being observed and perform a better diagnosis of the situation.

Tomosynthesis is also an x-ray based technique which beams the patient’s breast with radiation from

different angles so that several images can be taken from different perspectives (Projection View (PV)

images).

9

Page 28: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Figure 1.6: Angular capability of Siemens Model. This particular model creates up to 25 images. Source: To-mossıntese no Diagnostico do Cancro da Mama [15]

The first part of this process is similar to a mammography exam: the patient’s breast is placed

between an electronic radiation detector and it is compressed in order to obtain several digital images

of the breast.

Afterwards, the radiation beamer swipes over the breast, with an angular window that can go from

[-7.5º ; +7.5º]4 up to [-25º ; +25º]5, as seen in figure 1.6. This technique, creates up to 25 PV images,

which means that it is possible to have up to an image for every 2 degrees. [15] [16]

After the data from the detector is obtained, there is a need to reconstruct the images into a 3D

volume. Such a process can be done through several algorithms such as Niklason algorithm, filtered

back projection or matrix inversion tomosynthesis [17] and it compiles a 3D volume with 40 to 70 2D

slices, where the slices are separated by approximately 1 mm. [8]

Figure 1.7: Example of a tomosynthesis exam where the relationship between the different slices can be seen.Source: Washington Radiology Associates [18]

4In Hologic Models5In Siemens Models

10

Page 29: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

It is based on this reconstruction that a radiologist will perform the diagnosis. Resorting to image

processing techniques, he or she can vary the contrast of the images to what suits him/her best and can

easily swipe through the different slices to find any anomalies and report them to any doctor of another

area of expertise.

1.4 Computer Aided Diagnosis

In order to improve diagnosis’ accuracy, a concept was introduced called Computer Aided Diagnosis.

CAD is the use of techniques and algorithms to show/indicate where abnormalities are located. In

breast imaging, CAD is built by gathering several mammographies and use them as a training set. By

fetching characteristics of these exams, regions suspicious of being lesions can be displayed to doctors

in order to help them detect abnormalities and to speed up the diagnosis process. This way, suspicious

lesions that doctors would not notice without CAD are enhanced and wrong diagnosis due to fatigue and

tiredness can be reduced.

In figure 1.8 is displayed the output of the CAD system built in this thesis, where it is possible to see

the contour of the breast (skin) and the lesion previously marked by a doctor in green and the suspicious

region detected by the CAD system in red.

Figure 1.8: CAD output displayed where the previously marked lesion can be seen in green and the CAD sugges-tion in red.

The several processes used to reach this output are explained throughout the next chapters. It is

important to enhance that CAD systems do not replace doctors, nor should their outputs be displayed to

patients, for an interpretation of these results should still take place by an expert.

11

Page 30: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

12

Page 31: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

2State-of-the-Art

Contents

2.1 Pre-processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

2.2 Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17

2.3 Feature Extraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

2.4 Learning and Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

2.5 Summary of State-of-the-Art . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

13

Page 32: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

14

Page 33: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Introduction to State-of-the-Art

The two breast cancer screening techniques presented earlier, mammography in section 1.3.1 and

Digital Breast Tomosynthesis (DBT) in section 1.3.2 are conceptually different but use the same acqui-

sition techniques. Given this, when building a CAD for either, one should consider methods applied in

the other, for some of the issues and procedures to solve them might be the same.

This chapter summarizes methods, procedures and algorithms used by several authors throughout

the last years. It is divided into 4 sections: Pre-processing, Segmentation, Feature Extraction and

Learning and Classification, where each presents the steps taken for mass detection in either one of the

screening techniques. Each section includes a small explanation of its purpose and descriptions of what

the authors did in their studies.

2.1 Pre-processing

The pre-processing stage is the earliest one. This increases its importance for any alterations in

the data will influence all following stages. Many techniques presented in this section can be adapted

from mammography to DBT (and vice-versa) within two conceptions: either apply 2D procedures to

mammography and to the layers of the DBT exam, or scale (if possible) the implementation of a 2D

algorithm in a 3D environment. The main goal of pre-processing is to perform noise reduction, image

normalization, object enhancement, background-foreground separation among others.

Jun Wei et al. in [19] developed a CAD system for mass detection in Full-Field Digital Mammography

(FFDM). In this study, the authors start by applying Otsu’s method to compute the separation between

the background and the foreground (breast) of the image. This is done by labelling the several regions

of high intensity of the binary image and consider the region with the largest area to be the breast. An

enhancement technique used by R. Alam et al. in [20] is to use a linear transformation filter to enhance

the low and the high grey level areas of the image. The same parameters used in this technique are

applied to all images, independently of their initial contrast. This enhancement proved to increase the

sensitivity of the system and helps to visualize the elements.

When pre-processing DBTs, H.P. Chan et al. in [21] use a median filter. This is a clean-up technique

that performs noise removal whilst preserving edges. Another filtering technique is the bilateral filter. In

this filter, the value of each pixel is replaced by a weighted average of the surrounding pixels. The weights

can be computed via a Gaussian function and depend on the Euclidean distance and the intensity

values. To compute a bilateral filter, one can use the expression presented in 2.1.

15

Page 34: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

IBL(x) =∑xi∈Ω

I(xi)× fr(I(xi)− I(x))× gs(|xi − x|) (2.1)

where

• IBL(x) is the filtered image

• I(x) is the original image

• x is the pixel being considered

• xi are the surrounding pixels

• Ω is the window of analysis

• fr is the kernel used to smooth differences in intensities

• gs is the spatial kernel for smoothing differences in coordinates

An example of a bilateral filter is shown in figure 2.1

Figure 2.1: Example of a bilateral filter in which the centre in question is in a high intensity area but close to theborder. In a) it is possible to see a sharp transition between a dark and a bright region. In b), an exampleof a bilateral filter where one can see that the pixels with the same intensity as the centre, have higherweights. In c) the result of the bilateral filter applied to the transition shown in a). Source: [22]

S. Vedantham et al. in [6] and R. Thyagarajan et al. in [23] use a bilateral filter to remove ”X”-shaped

out-of-slices artifacts that happen due to the reconstruction process in DBTs. In order to perform edge

enhancement, these authors used an unsharp masking technique along the x-y plane. This masking

technique increases high frequency noise, so the authors used an anisotropic diffusion to decrease

such effect. This diffusion, used for noise regularization, minimizes the effect of noise while at the

same time preserves the edges (important for the segmentation stage). A last pre-processing technique

these authors suggest is background correction via the user’s input. By clicking on the location of the

Volume of Interest (VOI) the user wants to analyse (a potential lesion), a normalization is performed.

The expression to compute this normalization is presented in equation 2.2.

I(x, y) =I(x, y)− IminImax − Imin

, I ∈ V OI (2.2)

This normalization is done to correct the differences in the average of the layers’ intensity which can

be problematic when performing segmentation.

16

Page 35: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

X. Qin et al. in [24] use 4 different approaches to increase image quality in DBTs. First, an angular

constrained bilateral filter is needed due to limited angle imaging which causes out-of-slice artefacts.

Second, due to intensity based classification later in the process, cupping artefacts need to be corrected,

since the intensity values differ from the centre of the breast to the margins (in the reconstructed image).

The cupping artefacts are corrected using a non-parametric coarse-to-fine approach with an entropy

related cost function. Third, histogram equalization is used to enhance details and local contrast. Fourth,

authors also use a bilateral filter to remove noise in the x-y plane.

I. Reiser et al. in [25] study DBTs and perform two pre-processing techniques. First, a logarithmic

transform is applied to the PV images to enhance the contrast. Second, morphological denoising is

done to remove single and double pixel noise peaks, leaving three pixel regions (or larger) intact.

2.2 Segmentation

Segmentation is the name given to the act of taking out/identifying an object within an image. The

goal of this second step is to identify Regions of Interest (ROIs)1. Some segmentation techniques are

not object dependent and aim to segment any element found in the image. The techniques presented in

the following studies may derive from others, but are directed to find regions suspicious of being masses

in mammograms or DBTs.

M. R. Hejazi et al. in [26] built a CAD system for mass detection in mammographies where the

authors start by marking foreground objects using morphological operations such as erosion and dilation.

Later, regional maximums of intensity are considered the centre of the suspicious object and, from this

centre, inner and outer borders of the object are found by considering other pixels to be a part of the

same object if the difference in the average of the histograms (including and excluding the pixel) did not

overcome a certain threshold. This method outputs a region suspicious of being a mass.

Another segmentation technique is called Gradient Field Analysis (GFA). This analysis considers

areas of high convergence of the gradient to be the best mass candidates. It can be performed either

in 2D images or 3D Volumes. Jun Wei et al. in [19] use a gradient analysis technique to find the centre

and borders of ROIs. The technique was introduced by H. Kobakate et al. in [27] and it is called a COIN

filter (in its simpler version) or an iris filter (in a complex version) for it simulates the same adaptation

capacity of the human eye by calculating the convergence of gradient around each pixel, as displayed in

figure 2.2. This technique is used because it starts by searching for possible centres of ROIs and further

it finds the borders of the object whose centre was previously found. In 2008, H-P Chan et al. in [21]

proposed to use 3D GFA to segment masses in DBT. The approach suggested is to use GFA to pre-

1In the scope of this work, ROI is any region suspicious of being a mass.

17

Page 36: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

screen the mass candidates. Choosing a maximum of 30 sites with the highest gradient convergence,

3D GFA is used to select the mass candidates which will be further analysed. These authors suggest

the use of the same method to perform segmentation of mass candidates in 2D PV images. B. E.

Caroline et al. in [8] suggest the use of GFA to segment both 2D PV images and 3D volumes. The

authors suggest performing segmentation with a very high sensitivity rate, which means that the number

of false positives becomes also very high, most likely. This does not become a problem since these false

positives should be eliminated further ahead. The approach suggested in 2D is to detect the centre of

the ROI by computing the value of convergence of the gradient into a certain pixel. This value is called

Radial Gradient Index (RGI). The lesion candidates will be the ROIs centred in the pixels which exceed

a certain RGI value. The authors started by using a threshold of 10 %, which means selecting the top

10% pixel values (of RGI value) as centres of suspicious mass candidates. In 3D volumes, the detection

is similar to the description in 2D.

Figure 2.2: Example of a COIN filter being computed for a circle. P1 and P2 represent two pixels of interest beingconsidered for the centre. This technique considers the angle θ between the gradient vector in a givenpoint (k, l) and a straight line uniting (k, l) to the centre of interest. If that angle is close to zero (itscosine close to one), it means the gradient in (k, l) is pointing towards the centre of interest. In thiscase, P1 is chosen as the centre, for it was the point with most gradients point towards it. This iscomputed by summing the cosines of all the cases analysed. The norms and directions of the vectorsare not in scale and are simply illustrative.

R. Alam et al. use an adaptive local thresholding to find masses for these are denser than the

surrounding tissue, which means they will be a locally bright spot. This adaptive technique uses two

windows (a large and a small one) centred in the same pixel. It considers the pixels inside the small

window to be a mass if the difference between the intensity of small window and the large window to

overcome a certain threshold. This threshold is image dependent, which originates the name of the

technique.

A different approach which is used by several authors to segment lesions, or simply segment different

18

Page 37: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

tissues, are clustering techniques. This segmentation is performed by grouping pixels that have the

same characteristics together. For example, S. Vedantham et al. in [6] use Fuzzy C-Means (FCM) for

lesion segmentation by clustering pixels with similar grey levels. The authors use FCM with two clusters

in order to segment the lesion from the background tissue. After FCM, a binary mask is created with the

cluster label. Then, several morphological operations are applied to this mask, such as area filtering,

binary opening followed by binary closing to remove salt and pepper noise and also binary erosion and

binary dilation to perform mask smoothing. This is done to respond to the fact that FCM may create

isolated pixels, hence morphological operations are required.

Another clustering technique is Expectation Maximization (EM). It was used by R. Thyagarajan and

Dr. S. Murugavalli in [23]. These authors used FCM or EM due to their robustness and efficiency

and because these techniques are effective in segmenting noisy reconstructed images and incomplete

datasets. After using FCM, the authors also performed morphological operations for the same reasons

presented in [6]. Segmentation is performed using EM for maximizing and enhancing the affected areas

and is used to distinguish the anatomical structures of the breast.

2.3 Feature Extraction

After performing segmentation, some characteristics must be extracted from the regions or volumes

segmented. These characteristics are called ”features”. For CAD systems for mammographies or DBTs,

one can divide the features into 3 big groups:

• Morphological Features

• Grey Level Features

• Texture Features

This section is divided into 3 parts, in order to present the features used by different authors in their

studies, by group.

The first ones, morphological features, are all the features related to space, physical and shape

context. For mammography, M. R. Hejazi in [26] extracts 4 morphological features: Area, Shape, Edge

Distance Variation and Spread Pattern. Area is simply the total number of pixels in a segmented ROI.

Shape is extracted for discriminating circular masses from non-circular ones. This feature is calculated

by computing the distance between the centre of a region and the points at the border. Edge Distance

Variation indicates the difference between intensity levels at the border because usually tumours are

well-defined structures. This feature is computed by subtracting the intensity level of the pixel at the

border by the local mean. In this case, the computation of the local mean is not specified. Spread

19

Page 38: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Pattern analysis the inner and outer border, both previously computed, which indicates if a tumour is

benign (encapsulated and does not spread) or malignant (risk of spreading cancerous cells to other

parts of the body). D. Saraswathi et al. in [28] and [29] write about features extracted in mammograms

and in DBT. As morphological features, Area, Centroid Position, Orientation, Perimeter, Euler Number,

Projection, Thinness and Aspect Ratio are extracted.

For DBT, B. E. Caroline et al. in [8], suggest the extraction of the Candidate Ratio Index. This index

is the number of times a certain region is presented in the same position throughout all the PV images.

The computation of this index is shown in figure 2.3. As it can be seen, structure 1 is a mass so when

analysing PV images from the source in both positions 2 and 10, it will appear in the same location.

The same does not happen for structures 2 and 3, which are dense tissue. From position 2, structure

2 overlays structure 3, which creates the illusion of a denser tissue, possibly a mass. This hypothesis

is eliminated with the PV image from the source in position 10, where these two dense tissues do

not overlay each other. Therefore, this feature holds information on the morphology of the suspicious

structure and it is used to eliminate false positives that may be caused by overlaying tissue. H-P Chan

et al. in [21] extracted morphological features such as: Volume (in terms of number of voxels in the

object), Volume difference before and after 3D morphological opening, the surface area, the maximum

perimeter of the segmented object throughout all the slices and the compactness (percentage overlap

between the object with a sphere of the same volume, both centred in the same centroid).

Figure 2.3: Representation of the computation of the CRI. Structure 1 is a mass and structures 2 and 3 are densetissue. With the Projection View image obtained from source position 2, the combination of structures2 and 3 would appear to be a mass when in fact it would just be a representation of overlaying densetissue. Source: B. E. Caroline et al. in [8]

Grey level features are all the features related to the intensity of the region being analysed. D.

20

Page 39: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Saraswathi et al. in [28] and [29] extract the Mean, the Standard Deviation, Skewness, Energy and

Entropy from the ROI’s histogram. X. Qin et al. in [24] extracted 8 grey level features for further tissue

classification: Skewness, Contrast, Correlation, Entropy, Sum Entropy, Difference Entropy, Information

Measures of Correlation and Mean Intensity of each region. H-P Chan et al. in [21] also extracted grey

level features from the object’s histogram such as, maximum, minimum and average grey level intensity

of the 3D object, contrast between the object and its surroundings, Skewness, Kurtosis, Energy and

Entropy.

Texture features do not have a direct definition. The concept of texture is very vague and some

authors have tried to define texture qualitatively [30] but no definition ever stood out. This means that

texture features can be all the features that have information regarding the environment of the segmented

image. The periodicity of the image, the total gradient intensity, etc, can be considered as texture

features. Jun Wei et al. presented in [19] a way of using the wavelet transform to better extract texture

features from mammograms by performing an analysis on the Spatial Grey Level Dependence (SGLD)

in several frequency bands. This matrix contains informations regarding how often do pixels with the

same intensity repeat itself, which, when analysed locally, contain information on the irregularity or

homogeneity of a certain ROI. H-P Chan et al. in [21] extract two types of texture features: Run

Length Statistics features and SGLD matrices’ features. Of the first type, short run emphasis, long run

emphasis, grey level non-uniformity, run length non-uniformity and run percentage are extracted from

the Sobel gradient image. Run Length Statistics (RLS) features are extracted for they are effective in

differentiating spiculated from non-spiculated objects. When a structure is spiculated, it means that it

has spikes or points on its surface, which may be an indicator of malignancy [14]. In this study, 13 SGLD

features are extracted from each ROI. The final SGLD feature vector contains the average between

all the slices of each of the 13 features. D. H. Kim et al. in [31] propose to extract texture features

from both breasts and analyse their dissimilarities. In order to characterize the bilateral asymmetry

between VOIs on the left and on the right breast, three features are extracted: Histogram Dissimilarity,

Sum of Squared Differences (SSD) and Dissimilarity of Single View. Histogram dissimilarity has a low

value when VOIs are similar, so it is expected to output high value for masses. Given that masses are

asymmetric structures, the SSD between two VOIs should have higher values if one of the VOI contains

a mass. The Dissimilarity of Single-View is the absolute difference between a feature value of a VOI and

the feature value of that VOI’s version on the other breast. Many of these features are suppose to output

zero in a case where a VOI would become a False Positive (FP) for it is expectable that FPs have similar

tissue aspect in both the left and right breast.

21

Page 40: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

2.4 Learning and Classification

Learning and Classification is used in order to identify whether a region is cancerous or not. Most

CAD systems used for breast cancer are supervised, which means a labelled dataset is required, where

one has several patterns with the corresponding (correct) label. As exemplified in figure 2.4, one can

identify to which class a region belongs, giving it a label (classifying it). Naturally, these methods have a

certain error associated to them, which is what must be reduced.

Figure 2.4: Representation of a labelled dataset and a new pattern, waiting to be classified.

M. R. Hejazi et al. in [26] propose a mass detection method for mammograms, where several ma-

chine learning techniques are combined using explicit knowledge from experts in the field. The authors

start by ruling out patterns (consider them healthy tissue) if the area of the ROI is not between a max-

imum and a minimum threshold. The edge intensity variation and shape of the patterns that overcome

this test are used as features of two different Multi-layer Perceptron (MLP) that output the characteristics

of the tumour in a categorical way 2 3. This categorization is then used in a rule-based classifier built

by experts where situations such as ”if a tumour has only one segment circular and both segments are

clearly defined, then it is a malignant tumour” are used to make the final decision of the whole CAD

system. The sample used by these authors is a small one, being the authors able to identify 14 out of

15 benign tumours and 4 out of 5 malignant ones.

Nicholas Petrick et al. in [32] use the features extracted as an input to a Linear Discriminant Analysis

(LDA) classifier. These authors only separate masses from healthy tissue, using texture features and

this classifier, at a True Positive (TP) rate of 90%, outputs 3.77 FPs per image.

2For shape: inner and outer segments circular, only one circular or none.3For edge intensity variation: both segments clearly defined, only one or neither.

22

Page 41: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

H-P Chan et al. in [21] implement an ensemble classifier for classification of masses in DBT. Using

both 2D techniques to identify lesions in PV images and 3D techniques to identify lesions in 3D volumes,

the study suggests a combined approach to classify masses and reduce the number of FPs. The

authors use a dataset of 100 DBT cases and use only MLO DBT views. In 3D, in order to classify which

VOIs contained masses or not, the authors use an LDA based classifier, using the mass candidates as

patterns, and perform a leave-one-out cross-validation technique to train the classifier and to perform a

stepwise feature selection. This way, features can be selected from the feature pool and LDA weights

can be estimated. Alongside, in order to detect masses in 2D PV images, a rather similar process

occurs. Using only the ROIs considered as candidates after the segmentation step, a leave-one-out

method (with these ROIs as patterns) is implemented in order to build an LDA classifier. Geometry is

used to know the 3D location of the already classified 2D PV images. In order to obtain a classification

whether a candidate is a mass or not, the labels obtained from 3D VOIs and 2D ROIs are matched using

the following criteria: if the centroid of the 3D VOI classified as positive falls within the area classified

as a mass in 2D as well, the global system classifies that VOI as a positive. If the contrary happens,

which means, if a ROI is classified as positive and its centroid falls within a VOI which is also classified

as positive, the system also classifies this region as a mass. Only in the case where an object was

classified as positive in one of the approaches but is does not match any corresponding object from

the other approach, the object is not considered as a mass. At a sensitivity of 90 %, the presented

implementation obtained 4.05, 3.40 and 2.04 false positives per breast, for the 2D, 3D and combined

approach, respectively. This proved that the combined approach reduced the number of false positives

per breast.

D. H. Kim et al. in [31] use a dataset composed of 160 reconstructed volumes from the two views (CC

and MLO) from 40 patients, to perform mass detection. In this dataset, 54% of the volumes are biopsy-

proven to contain a malignant mass. For classification, a Support Vector Machine (SVM) algorithm with

radial basis function is used to distinguish masses from FPs. 5-fold cross-validation is applied to find

the optimal parameters of the SVM. The authors perform three experiments: one in which only single-

view features are used, another where bilateral features are used and a last one where single-view and

bilateral-view features from VOIs previously selected as mass candidates are used. The Area Under

the Curve (AUC) for each of the experiments is 0.809, 0.785 and 0.826 respectively. The FP per breast

are 1.99, 2.21 and 1.67 respectively. As it can be seen by the results, the use of both single-view and

bilateral features outputs the best results in terms of AUC and eliminating False Positives.

23

Page 42: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

2.5 Summary of State-of-the-Art

Table 2.1 presents four types of studies:

1. CAD systems for segmenting masses in mammograms.

2. CAD systems for the classification of masses in mammograms.

3. CAD systems for segmenting masses in DBT.

4. CAD systems for the classification of masses in DBT.

As one can see, the false positives that occur are still an existing problem even in the most recent

studies. On the other hand, very often, the patient’s dataset size is not statistically significant in many of

the studies.

24

Page 43: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Tabl

e2.

1:Ta

ble

pres

entin

gch

rono

logi

cally

seve

rali

mpo

rtan

tcon

trib

utio

nsin

the

field

ofC

AD

ofbr

east

canc

erin

mam

mog

raph

yan

dD

BT

inth

epa

stre

cent

year

s

Dat

eC

onf.

/M

aga-

zine

Aut

hor

Type

of stud

yA

ppro

ach

Seg

men

t.A

lgor

ithm

Lear

ning

Alg

orith

mFe

atur

esex

trac

ted

Dat

aset

size a

Sen

s.FP pe

rB

reas

t

2005

-M

.R.H

ejaz

iet

al.[

26]

2D

etec

tion

oftu

mou

rsvi

atw

ose

gmen

ts

Reg

iona

lm

axim

umof

inte

nsity

MLP

and

Rul

eB

ased

Cla

ssifi

ers

Are

a,S

hape

and

Edg

eIn

tens

ityVa

riatio

n80

0,8

-

2005

Med

ical

Phy

sics

(MP

)Ju

nW

eiat

al.[

19]

2U

sefro

mdi

ffere

ntfre

quen

cyba

nds

GFA

LDA

Mor

phol

ogic

al&

SG

LD20

00,

91,

75

2008

MP

HP

Cha

nat

al.[

21]

42D

imag

esan

d3D

VO

IG

FALD

AM

orph

olog

ical

,G

rey

Leve

l,Te

xtur

efe

atur

es10

00,

851.

63

2009

IJB

BE

R.A

lam

etal

.[20

]1

Hyb

ridS

egm

enta

tion

syst

em

ATan

dFu

zzy

Ent

ropy

Fuzz

yE

ntro

pyTe

xtur

efe

atur

es78

0,9

-

2011

IEE

ES

.Ved

anth

amat

al.[

6]3

Seg

men

tatio

nof

the

entir

ebr

east

FCM

-gr

eyS

cale

Leve

l98

--

2012

IEE

EB

.E.C

arol

ine

atal

.[8]

42D

imag

esan

d3D

VO

IG

FALD

A&

SV

M

His

togr

am,

Mor

phol

ogic

al,

Gre

yle

vel,

Text

ure

-0,

91.

61

2012

IEE

ER

.Thy

agar

ajan

etal

.[23

]3

Seg

men

tatio

nof

the

brea

stba

sed

oncl

uste

rsE

M,F

CM

-gr

eyLe

vela

ndTe

xtur

e4

0,98

-

2014

SP

IEX

.Qin

atal

.[24

]3

Div

ide

the

two

type

oftis

sue:

Adi

pose

vsG

land

ular

tissu

e

Bila

tera

lfil

terin

g,FC

MFC

MTe

xtur

efe

atur

es5

0,92

-

2015

IEE

E-

ICIP

D.H

.Kim

atal

.[31

]4

Sim

ilarit

yof

both

brea

sts

in3D

VO

I-

SV

Mw

ithR

BF

His

tDis

sim

ilarit

y,S

SD

,Dis

sim

ilarit

yof

Sin

gle

Vie

w40

0,85

1,67

aN

umbe

rofp

atie

nts

25

Page 44: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

26

Page 45: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

3Pre-Processing and Segmentation

Contents

3.1 Pre-Processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

3.2 Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

27

Page 46: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

28

Page 47: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Overview

This work has the general architecture presented in figure 3.1. It is divided into 4 steps, as detailed.

Each section of this chapter and the next explains one of the steps in the following order.

• Pre-Processing: techniques to clean noise and enhance suspicious regions are used to improve

ROI detection and lesion classification.

• Segmentation: a searching algorithm is implemented to locate such ROIs within the image.

• Feature Extraction: specific characteristics of regions are extracted to use as input of the classi-

fiers.

• Classification algorithms: two algorithms are set up to classify classify the detected lesions.

Figure 3.1: Scheme of the 4 steps approach applied in this thesis

3.1 Pre-Processing

The pre-processing step is divided into two parts: identifying the breast area and cleaning the image

for better ROI detection (segmentation).

To identify the breast, Otsu’ method was used [19]. This method separates the background pixels

from the foreground (breast). It assumes that there are two-classes of pixels in the image (background

and breast) and exhaustively searches for the intensity value which separates the two classes better.

This means that Otsu’s method tries every value between 0 and 255 and chooses the one which min-

imizes the variance of the set of pixels below that value and the set of pixels above that value. The

conceptual idea is to identify that there are two modes in the image’s histogram, and find the intensity

level that separates the two. In this case, the breast is considered to be the larger region with higher

intensity level. In figures 3.2(a) and 3.2(b), background and foreground modes can be visually identified:

background being the bin most to the left (dark pixels), in intensity value 0, and foreground being the

maximum located in the bin centred in intensity level 25.

29

Page 48: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Several local maxima can be seen in the histogram. To solve this, there is the need of applying a

Gaussian filter to the whole image, as seen in figures 3.3(a) and 3.3(b). This process blurs the image

which brightens possible dark zones within the breast. The result is shown in figure 3.3(a), where it can

be seen that the breast becomes a generically bright spot.

(a) Digitalized Mammogram (b) Histogram of digitalized mammogram

Figure 3.2: Example of a a) digitalized mammography from DDSM and b) its histogram. As explained before, it canbe seen that even within the breast, darker areas may exist

30

Page 49: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

(a) Blurred Digitalized Mam-mogram

(b) Histogram of digitalized mammogram and Otsu’smethod border

Figure 3.3: Example of a) mammography and b) its histogram, where a Gaussian filter with a standard deviation of50 is applied. In this image, it can be seen that the breast area becomes a generically high intensityregion. Histogram displays Otsu’s method’s threshold in red at intensity level of 3. Pixels with intensitylevel from 0 to 3 belong to the background, pixels with intensity levels from 3 to 255 belong to theforeground. It is visually possible to realize that the transitions smooth due to the filter, hence, it iseasier to find the local maxima.

31

Page 50: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

(a) Binary mask with all components (b) Binary mask only with biggest com-ponent

Figure 3.4: Binary image obtained from dividing the mammography presented in 3.2(a) into background and fore-ground. Notice on the top right corner of a) that other elements were wrongly identified as foreground.Hence the need to withdraw only the larger object as seen in b).

Figure 3.5: Resultant image of separating breast (foreground) from background using Otsu’s method. The greenline represents the border of the mask computed from this breast. This process is important for manyof the methods following this one will be applied only to pixels within the mask.

32

Page 51: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Next, a filter will be applied in order to reduce noise. The bilateral filter [24] was chosen due to its

ability to reduce noise without losing information of the edges, which is important for the following step

(Segmentation). This goal is achieved by combining two Gaussian filters in order for two criteria per pixel

to be considered: intensity and spatial distance. These criteria are used to compute the weight of every

pixel in a given neighbourhood of the pixel being analysed. In an image processing context, the concept

of a filter is to perform a sum of products within a certain window. In this case, the filter size is 81 × 81

pixels. This means that every pixel 40-pixels away in any direction of the one of interest, is going to be

taken into account to compute its filtered value.

Let’s consider image I and filter f , where f has size 3 × 3, being its centre in f(1, 1). Considering

the pixel of interest to be (cx, cy), then the resultant image If , is the sum of products of the pixels of I

with filter f , centred in pixel (cx, cy), which is presented, generically, in equation 3.1.

If (cx, cy) =

1∑x=−1

1∑y=−1

I(cx − x, cy − y) × f(1− x, 1− y) (3.1)

The concept of the bilateral filter used in this work is to combine two filters into one. Let’s consider

filter f to be related to spatial distances, filter g to be related to intensity distances and filter r to be the

resultant filter, with the following expressions:

f(cx, cy, x, y) =e−|(cx,cy)−(x,y)|

2σ2s

, (x, y) ∈ Ω (3.2)

g(cx, cy, x, y) =e−(I(cx,cy)−I(x,y))2

2σ2I

, (x, y) ∈ Ω (3.3)

r(cx, cy, x, y) = f(cx, cy, x, y)× g(cx, cy, x, y), (x, y) ∈ Ω (3.4)

where σs and σI are the standard deviation to consider in spatial and intensity dimensions, respec-

tively and Ω is the window of analysis. In this work, σs = 20 and σI = 0.121. The illustration of the filters

can be seen in figure 3.6, 3.7 and 3.8, respectively.

After the resultant bilateral filter r is computed, the filtered image IBF is achieved through equation

3.5.

IBF (cx, cy) =∑x,∈Ω

∑y∈Ω

I(x, y)× r(cx, cy, x, y) (3.5)

1Fraction of the maximum intensity value of the entire image

33

Page 52: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

(a) 3D representation (b) X-y plane representation

Figure 3.6: 2D spatial filter represented in a) 3D and b) from the x-y plane. In this figure, it is possible to see thatthe further way a pixel is of the pixel of interest (center), the less its weight is going to be.

(a) 3D representation (b) X-y plane representation

Figure 3.7: 2D intensity filter representation in a) 3D and b) from the x-y plane. The pixels with the highest weightvalue represented in this filter are the ones with the intensity value closest to the value of the pixel ofinterest.

34

Page 53: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

(a) 3D representation (b)

Figure 3.8: 2D bilateral filter representation in a) 3D and b) in x-y plane. Combination of filters represented infigures 3.6 and 3.7, where it can be seen that pixels closer to the pixel of interest in terms of spatialdistance and intensity level value will have higher weights.

3.2 Segmentation

The segmentation step is done in order to identify ROIs suspicious of being a mass. This will allow

the CAD system to extract information of regions suspicious of being lesions, instead of performing a

greedy search on the entire image, which can become unfeasible in terms of computational time and

can output many FPs.

In order to perform segmentation, the Sliding Band Filter (SBF) is used as implemented by Tiago

Esteves et al. in [33]. In this study the authors use this technique to find cells in microscopic fluorescent

images, where the cells’ pixels (foreground) have a higher intensity than the background (tissue). These

are the same conditions of the masses in the mammography images used in this thesis, as it can be

seen in figure 3.9. The authors use this filter for its ability to handle illumination variations, noise and low

contrast.

35

Page 54: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Figure 3.9: Comparison between image from the database used in this thesis (to the left) and microscopic fluores-cent image (to the right) segmented by the authors in [33].

SBF is part of a larger group of filters called Local Convergence Filters (LCF). These filters perform

Gradient Field Analysis (GFA) which considers gradient orientation instead of gradient magnitude. This

is an important characteristic for segmenting ROIs in this work, for the filter becomes independent of

the gradient magnitude, and therefore, independent of image properties, such as maximum or minimum

image intensity level. In order to understand this concept, it is important to understand the concept of

image gradient.

The gradient of an image is used in order to know if there are local changes in the pixel intensity.

Assuming a grey scale image, by computing the gradient of a pixel, it is possible to know how big such a

change is (magnitude) and which is the direction of the change (orientation). This is exemplified in figure

3.10(a) and 3.10(b).

(a) Grey-scale image and its gradients (b) Matlab quiver representation

Figure 3.10: Gradient representation in a a) black and white image and b) its quiver representation from Matlab

To compute the gradient, let’s consider image I and 5I as the gradient of image I [34]. Since an

36

Page 55: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

image gradient will have two directions, one can say that:

5 I =

[gxgy

]=

[ ∂I∂x∂I∂y

](3.6)

Following this, one computes the magnitude and direction of the gradient [30]. This can be done

through:

5 Imag =√g2x + g2

y (3.7)

5 Idir = tan−1

(gxgy

)(3.8)

A more graphical way to represent the gradient, is to think about the pixel’s neighbourhood. As

represented in figure 3.11, a simple way to compute the gradient is to analyse the difference between

the pixel above and the pixel bellow, and the difference between the pixel to left and the pixel to the

right. These differences are gy and gx respectively. The computation of the gradient using this method

is called the Intermediate differences gradient [35].

Figure 3.11: Graphical representation of the gradient of a pixel computed using Intermediate Differences gradient

The main goal of SBF is to compute the degree of convergence for each pixel. This measurement

indicates the convergence of the gradient vectors towards a certain pixel (pixel of interest). In other

words, it analyses the surrounding area of the pixel of interest and computes ”how many” gradients point

towards that same pixel. The pixel is considered to be the centre of a ROI if its degree of convergence is a

regional maximum. There are several variations of LCFs and the SBF used in this work is a combination

of 3 of them: COIN filter, iris filter and adaptive ring filter.

As presented in section 2.2, SBF’s basic principle is similar to the one presented by H. Kobakate et

al. in [27] (the COIN filter). It starts by computing the gradient for each pixel in a certain window, as

represented in figure 3.12a, followed by computing the line segment which unites each pixel with the

pixel of interest, as represented in figure 3.12b. The final goal is to compute the cosine of the angle

37

Page 56: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

between the gradient vector and the line segment, hence computing ”how much” each gradient points

towards the pixel of interest. Let us call this value CI as presented in equation 3.9 where α represents

the angle between the gradient vector and the line segment and (x, y) are the spatial coordinates of a

pixel.

CI(xi, yi) = cos(α(xi,yi)) (3.9)

Figure 3.12a: Gradient of an image. Each arrow repre-sents the gradient vector in each pixel ofthe surrounding area of the pixel of inter-est.

Figure 3.12b: Line segments uniting each pixel with thepixel of interest.

Figure 3.13: Fusion of the gradient of each pixel and the line segment uniting each pixel with the pixel of interest.Graphical representation of the angle used to compute the degree of convergence.

To achieve the degree of convergence, a certain number N of orientation points is determined, from

which the maximum value of CI is computed. As shown in figure 3.14, the orientation points are spread

38

Page 57: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

through the entire radial space around the pixel of interest. SBF uses one of COIN filter’s property which

is to find the maximum value of the sum of CI among all orientation points, as described in equation

3.10 and figure 3.14.

DCCOIN (P1) = max0<R<Rmax

1

N ∗R

N∑i=1

R∑r=0

CI(xi, yi), where x2i + y2

i < R2 (3.10)

Figure 3.14: Illustration of COIN filter being computed for an approximately circular lesion. This figure exemplifiesa case where the centre is correctly found, but the border is not.

This method performs well when locating centres of regions, but it assumes a round shape (which

might not happen). For a better location of the border, the iris filter presents a solution where the border

of a non round shape can be found. The concept of this filter is to use the sum of the maximum value per

orientation point instead of the maximum value of the sum among all orientation points, as presented in

equation 3.11.

DCIRIS(P1) =1

N

N∑i=1

max0<R≤Rmax

[1

R

R∑r=0

CI(xi, yi),

]where x2

i + y2i < R2 (3.11)

39

Page 58: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Figure 3.15: Illustration of an iris filter being computed for the same approximately circular lesion as in figure 3.14.This figure exemplifies a case where both the centre and the border are found. In this illustration, only8 orientation points are used, hence, the border is not entirely found, although, if N → ∞ , it would.2

The last property used by SBF comes from the adaptive ring filter, a minimum and maximum radius

can be used to search for the border only in some specific locations in order to obtain better results. The

expression used to compute SBF is found in equation 3.12 and its concept is represented in figure 3.16.

DCIRIS(P1) =1

N

N∑i=1

maxRmin<R≤Rmax

[1

R−Rmin

R∑r=Rmin

CI(xi, yi),

]where x2

i + y2i < R2 (3.12)

Figure 3.16: Illustration of an adaptive ring filter being computed for the same approximately circular lesion as infigure 3.14 and 3.15. In this illustration, the Rmin and Rmax are used to limit the area in which theborder can be found.

2In reality, if N ≥ 2x+ 2y, being x, y the dimensions of the image, the border is correctly found.

40

Page 59: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

4Feature Extraction and Classification

Contents

4.1 Feature Extraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43

4.2 Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

41

Page 60: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

42

Page 61: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

4.1 Feature Extraction

In the feature extraction stage, information from each ROI is collected in order to classify the ROIs

into healthy tissue, malignant masses or benign masses. This chapter describes 3 types of features:

Morphological, Intensity and Texture based. These, on the contrary of the previous stage, are extracted

from the raw image, and not the pre-processed one. This happens because, although pre-processing

techniques are important for ROI identification, they interfere with the information present in the image,

hence, raw data is used.

4.1.1 Morphological Features

Morphological features are extracted so that the system has information on the shape characteristics

of the ROI being analysed. Given that the images used have different sizes, if no action is taken,

morphological features will get affected. For example, if a lesion has an area of x pixels and a different

lesion (in a different image) as an area of 2x pixels, it becomes impossible to know if the second lesions is

twice as large, or if is is just an image with twice the size. In order to oppose this issue, two actions can be

taken: either the images are resized to the same value or all extracted features are normalized to the size

of the image. In this work, the second option is chosen, for resizing images implies oversampling and

sub-sampling techniques that might affect the information extracted. Given this, all the measurements

are normalized to the image size.

One feature to extract is the location of the centre (output of the segmentation stage). The x and

y coordinates are normalized to the image length and height, in order not to be affected by the image

dimensions.

The following feature to extract is the maximum radius. If a lesion is perfectly circular, to extract only

the distance between the centre and any point of the border would be enough. In this case, one extracts

the euclidean distance between the centre and its farthest point. The computation of this feature is

presented in equation 4.1,

R = argmax(√

(Bx − cx)2 + (By − cy)2)

(Bx, By)(4.1)

where Bx and By are the coordinates of the pixels of the border (computed in the segmentation stage)

and cx and cy, the coordinates of the centre.

After the maximum radius, the number of pixels that are part of the border are extracted: the perime-

ter. For this computation, the number of pixels correspondent to the border, which are output by the

segmentation algorithm is not enough, for these are just a subset of such border, as seen in image 3.15.

This way, the perimeter is computed by summing up the distances between a pixel of the border sample

43

Page 62: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

and the following pixel, as presented in equation 4.2, where N is the number of orientation points.

P =

N∑i=1

(√(Bx(i)−Bx(i+ 1))2 + (Bx −Bx(i+ 1))2

)(4.2)

Next, the area is also to be extracted. Given that the region found is a closed region it is possible to

compute the amount of pixels inside this region. Like the previous features, it is normalized to the image

size.

Another feature to extract is the Surface Change (SC) after mathematical morphology is applied.

In this case, the difference between the original image and the image after opening of the ROI by a

disk with the size of 1/10th of the radius is extracted. The expression of the mathematical morphology

operation is presented in equations 4.3 and 4.4.

ROIMM = ROI disk = (ROI disk)⊕ disk (4.3)

SC = ROI− ROIMM (4.4)

”Opening”, in mathematical morphology terms, means to perform two operations between two im-

ages: erosion and dilatation [36]. In this thesis, opening is performed between a ROI and a structuring

element (disk). The first operation, dilatation, can be though as the union between the ROI and the disk

in every pixel. On the contrary, erosion can be considered the intersection between the two elements.

As an example, assuming image A as a square, figures 4.1(a) and 4.1(b) represent both operations.

(a) Dilation illustration (b) Erosion illustration

Figure 4.1: Mathematical morphology operation a) ”dilatation” and b) ”erosion” performed between image A, asquare, a the structuring element B, a disk. Source: Language and Media Processing Lab [36]

44

Page 63: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

The goal of this operation is to analyse how ”round” a certain lesion is. By applying mathematical

morphology opening, the more round a certain region is, the less effect an opening will have. Hence,

by extracting the difference between the original image and the transformed image, one has information

on the roundness of the ROI. Figures 4.2(a) trough 4.4(b) exemplify the different outcomes of extracting

this feature.

(a) Original image (b) Image where mathematical mor-phology was applied

Figure 4.2: Example of a round region where mathematical morphology opening was applied. At it can be visual-ized, the image does not differ from its original one.

(a) Original image (b) Image where mathematical mor-phology was applied

Figure 4.3: On the contrary of the region presented in figure 4.2, when opening is applied to an approximatelyround region, some changes can be observed already.

45

Page 64: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

(a) Original image (b) Image where mathematicalmorphology was applied

Figure 4.4: Opposing the two cases earlier, figures 4.2 and 4.3, when opening is applied to a spiculated region, asignificant difference can be seen.

The list of morphological features extracted is presented below:

• X and Y coordinates of the location of the center

• Maximum Radius

• Perimeter

• Area

• Surface Change

4.1.2 Grey Level Features

Grey level intensity features are also extracted. These features hold information regarding the inten-

sity values within the ROIs. Given that it is possible to define exactly which pixels are inside the ROI,

grey level features will be extracted for all pixels in it. Let’s consider a given ROI with N pixels.

The first features to extract are the mean (equation 4.5), the maximum and the minimum of the

intensity values of the ROI being analysed.

µ =1

N

N∑i=1

I(i), i ∈ ROI (4.5)

Another set of features, which also holds information regarding the ROI, are features from its his-

togram analysis: standard deviation, skewness and kurtosis. The standard deviation of intensity is the

feature illustrated in figure 4.5. Its computation is as defined in equation 4.6 and has the goal to measure

how spread the intensity levels are along the scale.

σ =

√√√√ 1

N − 1

N∑i=1

(I(i)− µ)2, i ∈ ROI (4.6)

46

Page 65: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Figure 4.5: Examples of histograms of ROIs with the same mean intensity value but with different levels of standarddeviation. On the left, an histogram with a small value of standard deviation. On the right, an histogramwith a standard deviation value 5 times higher than the one on the left.

The skewness of an histogram discriminates whether the histogram is equivalently or asymmetrically

distributed around the sample mean. The computation of skewness is presented in equation 4.7, being

E the expected value.

s =E(IROI − µ)3

σ3(4.7)

Similar to skewness, kurtosis is a feature that describes the shape of the histogram. It is a measure of

spikiness of the center of the histogram, allowing for its better description. Its computation process is

described in equation 4.8.

k =E(IROI − µ)4

σ4(4.8)

Another feature to extract relates to the connection between the intensity levels of the ROI and the

entire image: contrast. It is the ratio between the mean intensity value of the ROI (as presented in

equation 4.5) and the mean intensity value of the entire image.

The last intensity related single feature to extract is the energy, which is computed as presented in

equation 4.9.

Energy =

N∑i

I(i)2, i ∈ ROI (4.9)

Finally, a last set of features is extracted. A 10 bin histogram of the intensity levels of the ROI is

computed and the number of pixels present in each bin1 is considered as a feature. Given this, the list

of the grey level features to extract is presented below:1as described in section 4.1.1, the number of pixels is normalized to the image size

47

Page 66: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

• Mean Intensity Value

• Maximum Intensity Value

• Minimum Intensity Value

• Standard Deviation of Intensity Values

• Skewness

• Kurtosis

• Contrast

• Energy

• 10 Intensity Histogram bins

4.1.3 Texture Features

Regarding texture features, one single feature and a set of features are extracted.

The entropy is a statistical measure of randomness of intensity levels. This randomness indicates

if the ROI is smooth or wrinkled. This measurement is computed according to equation 4.10, where p

represents the fraction of pixels which have level of intensity l. In this work, the intensity levels are not

considered in bins but all separate (from 0 to 255).

Entropy = −255∑l=0

(pl ∗ log(pl)) (4.10)

The last set of features to extract is an analysis made to the histogram of the angles of the gradients.

These are called Histogram of Oriented Gradients (HOG) features [37] and hold information on the

ROI’s organization. For example, in figure 4.6a, one can see an image and its histogram. Given the

randomness of this image, the histogram bin levels are uniform, for there are gradient angles in every

direction. On the other hand, in figure 4.6b, there is an organized image with stripes and hence, its

histogram bin levels are not uniform any more. It can be seen that the gradients of the image are always

in the same direction and therefore, the majority of its pixels will be in the same bin (for they have the

48

Page 67: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

same gradient angle).

Figure 4.6a: Example of the angle’s histogram of an image. On the left, the histogram of an image with a randomtexture (presented on the right). By this histogram, it is possible to see that there are gradients pointingto all directions in the image on the left.

Figure 4.6b: Illustration of how an analysis on the histogram of the gradients’ angle of an image can be descriptiveof such image. In this case, on the right an image with stripes and on the left, its histogram. It canbe easily seen that, comparing to figure 4.6a, this histogram outputs more information regarding theorganization of its image.

49

Page 68: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

4.2 Classification

The CAD system implemented in this thesis for diagnosing breast cancer in mammography can

have its output vary between benign masses, malignant masses or healthy tissue. The classification

approach suggested in this thesis is to implement several classifier ensembles using different training

sets and different features. Each classifier is binary and outputs one of two options:

• Healthy tissue vs Mass

• Malignant mass vs Benign mass

This is done by building a 2-stage classifier ensemble, as seen in figure 4.7. The goal of the first

stage is to determine whether the ROI is a lesion or not, whereas the second stage determines its

malignancy. Given this, the first one distinguishes masses from healthy tissue and the other, benign

masses from malignant masses.

In both stages, the general approach is the same, but the datasets used are different. In the first

one, the dataset includes all patterns (healthy tissue, benign and malignant) and in the second one, only

benign and malignant are used. This division is made due to the final goal of the CAD system: to point

out the lesions to the doctor. This means that identifying a lesion is more important than saying whether

it is benign or malignant, given that if a lesion is detected, the doctor should be aware of it, regardless

of its malignancy. Due to these reasons, 2 stages are built since, this way, it is possible to parametrize

the CAD system to output the best results.

In each stage an ensemble of classifiers is used, where each base classifier in the ensemble uses

features from a different type. The output of the base classifiers is combined using combination tech-

niques (Majority Vote, Weighted Majority Vote, Naive Bayes Combination [38]). This way, the flow of

classification is compiled by a first stage divided into 3 classifiers (one for each type of feature: Morpho-

logical, Grey level and Texture) and a combination technique. Here, the patterns classified as healthy

tissue are considered non-lesions and leave the flow. The second stage which also contains 3 classifiers

and a combination technique, classifies the patterns into benign or malignant lesions. This second stage

performs worse, for lesions are more look a like among themselves than with healthy tissue.

In order to obtain the best results, two different types of classifiers (Decision Tree (DT) and AdaBoost)

are tested.

50

Page 69: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Figure 4.7: Representation of the classification flow used in this work. The features of the ROI are used as input for3 classifiers, whose outputs are combined and in which the ROI is classified as a Mass or as HealthyTissue. In case of a mass, the process moves along and the features are used as inputs to another 3classifiers, that, after its outputs are combined, define if the ROI is malignant lesion or a benign one.

4.2.1 Decision Tree

The decision tree used in this thesis uses the CART algorithm [39]. The general idea of a DT is to

use thresholds (for continuous features) or rules (for categorical features) to classify the patterns, as

shown in figure 4.8. A DT is a supervised method, which means it has to be trained beforehand. The

root of the tree (first node) is computed by selecting, from the feature pool, the rule which maximizes

the drop of impurity. After the node is selected, the same algorithm is applied to the subtree with the

corresponding subset of patterns, which were divided in the root. By doing this process, ones grows the

tree. One property of the CART algorithm is that only binary nodes are created.

The feature/rule used for each node is not arbitrarily chosen. Let’s assume an example: by using

the rule ”radius being larger than 50, the pattern is healthy tissue”, 54% of the patterns are wrongly

classified. On the other hand, by using ”mean intensity value larger than 180, the pattern is a mass”,

one sees that 70% of the classifications are correct. In this case, the second rule would be chosen, and

two subtrees would be built with two other nodes, and then four other, etc, etc. It is computationally heavy

(and unnecessary) to compute the classification of the tree in order to know which rule to choose. The

51

Page 70: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

measurement function used by the CART algorithm to define the rule is called Gini2, which computes

a measure called impurity. The expression to compute Gini is presented in equation 4.11, where S

represents the subset of patterns, C represents the number of classes (in this case, 2) and ps is the

fraction of patterns in each class.

Gini(S) = 1−C∑k=1

ps(k)2 (4.11)

In order to choose the rule, one computes the drop of impurity. For example, splitting sample S with

respect to feature A can be computed via equation 4.12:

Drop(S,A) = Gini(S)−∑

v∈values(A)

|Sv||S|

Gini(Sv) (4.12)

in which, Sv is the subset generated if the rule is applied.

By maximizing the drop of impurity, one can find out which rule to use when building the node. In

figure 4.9, the Gini function can be seen.

Figure 4.8: Example of a DT with three nodes and fourleafs. Figure 4.9: Gini function

A decision tree is grown until it reaches a certain stopping criteria. Among these criteria, one can

have:

• A maximum number of splits is achieved

• A certain threshold of drop value is achieved

• Each leaf corresponds to one class only

This growth has to be controlled, for over-fitting might occur when growing a decision tree. In order

to control it, the tree can be pruned.

2Other algorithms use other measurement functions

52

Page 71: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

In the current work, two classifiers were built using decision trees. The first experiment was a single

model hypothesis with a DT as a classifier, with a dataset containing all the patterns and three labels:

healthy tissue, benign lesions and malignant lesions.

Figure 4.10: Ilustration of the Single Model Decision Tree used in this work

A second experiment was done using a Decision Tree Ensemble in the same conditions: using all

the patterns and the three classes available, to test a simple version of the classifier ensemble and the

2 stage hypothesis. In this case, a total of 6 decision trees are trained: 3 for classifying the lesions into

healthy tissue or mass and 3 other for classifying lesions into benign masses and malignant ones. Each

model is trained with a specific set of features: morphological, intensity based or texture. A schema of

this classifier is presented in figure 4.11.

Figure 4.11: Schema of the classifier used in the Decision Tree Ensemble experiment.

53

Page 72: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

In order to understand better the order of the actions taken in this experiment, algorithm 1 presents

the pseudo-code used in it.

Algorithm 4.1: Pseudo-code of the algorithm used in the Decision Tree Ensemble experiment.Data: Dataset off all patterns (and their features)Result: Label vector C, the accuracy, true positive and true negative

for kfold = 1−→ 10 do%Stage 1

t(1)←− decision tree trained with morphological features of all patterns;t(2)←− decision tree trained with intensity features of all patterns;t(3)←− decision tree trained with texture features of all patterns;

CMat←− Compute confusion matrix for each of the trees t(1), t(2), t(3) (to be used by thefusion technique);

OutputTree(1)←− classify validation set with t(1);OutputTree(2)←− classify validation set with t(2);OutputTree(3)←− classify validation set with t(3);OutputsS1←− Naive Bayes fusion off all three outputs;

%Stage 2;t(4)←− decision tree trained with morphological features of masses;t(5)←− decision tree trained with intensity features of masses;t(6)←− decision tree trained with texture features of masses;

CMat←− Compute confusion matrix for each of the trees t(4), t(5), t(6) (to be used by thefusion technique);

if pattern classified as mass thenOutputTree(4)←− classify validation set with t(4);OutputTree(5)←− classify validation set with t(5);OutputTree(6)←− classify validation set with t(6);

OutputsS2←− Naive Bayes fusion off all three outputs;Final Output←− OutputsS2;

elseFinal Output←− OutputsS1;

Compute the accuracy for this fold using Final Output and Label.

Average the accuracy

54

Page 73: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

4.2.2 AdaBoost

The other method experimented in this work, AdaBoost [40] [41], is an ensemble classifier for binary

classification that trains T weak learners and combines their outputs into a strong classifier C. This

method can be considered a feature selection and learning algorithm, for each learner t chooses the

best feature (and its threshold) and performs a classification. Its name comes from the fact that it boosts

the classification performance of a single weak learner, in this case, single node decision trees, by

selecting the one with the highest classification rate (or lowest error). The learners are called ”weak”,

for they normally have a low classification rate, but the fusion of the outputs of the several learners

has proven to outcome robust results. Classifier C is a weighted combination of all weak learners that

outputs the classification solution. The added value of this ensemble method is that it accounts for the

output of learner t and, with it, adjusts the classification weights of each pattern, used in learner t + 1.

Figure 4.12 exemplifies this statement, where each learner picks a feature out of the feature pool.

Figure 4.12: Ilustration of the AdaBoost Classifier

In mathematical terms, each learner is a combination of patterns x, feature f , threshold θ and p, the

polarity indicating the direction of the inequality present in equation 4.13.

h(x, f, p, θ) =

1, if pf(x) < pθ

0, otherwise(4.13)

55

Page 74: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Adaboost is trained in an iterative way. During this process, two actions take place constantly:

• Computation of learner function h(t)

• Update of weights w(t)

The first one happens by minimizing the weighted classification error, present in equation 4.14 where

N is the total number of patterns, w(t)i is the classification weight for pattern i in learner t, di is the desired

output (pattern i’s correct label) and h(t) is the learner function.

ε(t) = minf,θ,p

N∑i=1

w(t)i |h

(t)(xi, f, θ, p)− di| (4.14)

It is worth noticing that the module portion of equation 4.14 is either 1 or 0. With this, h(t)(x, ft, θt, pt)

is chosen as learner t by being the one which minimizes function ε(t).

Figure 4.13 shows an example of the initial learner h(t), where the learner is represented by a border

that separates the patterns in the best way possible.

Figure 4.13: Illustration of the first learner being computed. In this example, all the patterns have the same weight(represented by the patterns size)

As mentioned before, in the beginning of each iteration, an update on the weights happens for w(t).

The weights are initialized according to the proportion of patterns from each class and are updated for

every t, as seen in equation 4.15,

w(t+1)i = w(t)β

(1− |h(t)

i −yi|)

(t) (4.15)

where β(t) = ε(t)

1−ε(t) .

In other words, this means that the weight in t + 1 is the same as the weight in t if the pattern is

incorrectly classified3 (|h(t)i − yi| = 1) or it is updated to a smaller relative value (dependent of ε(t)) as

represented in equation 4.16.

w(t+1)i =

wt, if |h(t)

i − yi| = 1 (incorrect classification)

wt × ε(t)

1−ε(t) , otherwise(4.16)

3Before the normalization step

56

Page 75: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Figure 4.14 and 4.15 illustrate the process of the weight update, when starting in figure 4.13, by

increasing the size of the pattern to exemplify the weight increment after normalization.

Figure 4.14: Illustration of the second learner of an AdaBoost. In this figure, the patterns previously miss-classifiedin figure 4.13 have a larger classification weight, as represented by their size.

Figure 4.15: Illustration of the third learner of the AdaBoost. As it can be seen, the patterns miss classified both infigure 4.13 or 4.14 have larger weights than the others.

In each step, after the computation, the weights are normalized. This means that if the patterns were

miss-classified, their weights increased.

N∑i=1

w(t+1)i = 1. (4.17)

After AdaBoost is trained, it is used for classifying the test and/or the validation set. Let’s consider Z

to be a set without labels, in which C is the label vector output by the AdaBoost. C is computed through

a fusion of the outputs of the several weak learners as presented in equation 4.18, where α(t) = log 1β(t) .

C(Zi) =

1, ifT∑t=1

α(t)h(t)(Zi) >12

T∑t=1

α(t)

0, otherwise

(4.18)

To sum up, the final classifier of the AdaBoost is a weighted combination of the outputs of the weak

learners, in which α(t) represent the weight associated to each one. Notice that the term to the right,12

T∑t=1

α(t), represents the situation where 50% of the learners classified pattern Zi as 1 and the other 50

57

Page 76: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

as 0, hence, if the term on the left overcomes this value, the pattern is classified as 1, otherwise, as 0.

The pseudo-code in algorithm 4.2 enhances the order of each step mentioned.

Algorithm 4.2: AdaBoostData: Given x to be the training dataset, y to be the label vector and Z to be unclassified dataResult: Vector C is the label vector of Z, output by the AdaBoost classifier

Initialization of weights according to the label, where wxc= 1

2Nc, where xc represents a pattern

of class c and Nc is the number of patterns in such a class;for t −→ T do

Normalize the weighs by w(t)i ←

w(t)i∑

jw

(t)j

;

Select h(t), f, θ and p by minimizing ε(t) = minf,θ,p

N∑i=1

w(t)i |h(xi, f, θ, p)− di|

Updates the weights with w(t+1)i = w(t)β

(1− |h(t)

i −yi|)

(t)

Compute the final weighted classification using

C(Zi) =

1, ifT∑t=1

α(t)h(t)(Zi, f(t), θ(t), p(t)) < 1

2

T∑t=1

α(t)

0, otherwise

In this thesis, a third experiment is done using an AdaBoost Ensemble. Opposite to the experiment

presented in section 4.2.1, instead of 3 classifiers, 300 are used (100 for each type), each using, for

training, 80% of the total training data obtained using bootstrap sampling with replacement [42]. The

schema for this experiment is as presented in figure 4.16.

Figure 4.16: Schema of the classifier used in the AdaBoost Ensemble experiment.

58

Page 77: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

In order for it to be compared with the other experiments, algorithm 3 presents the pseudo-code of

the AdaBoost Ensemble experiment.

Algorithm 4.3: Pseudo-code of the algorithm used in the AdaBoost Ensemble experiment.Data: Dataset off all patterns (and their features)Result: Label vector C, the accuracy, true positive and true negative

for kfold = 1−→ 10 do%Stage 1for loop = 1:3:300 do

train←− Sample 80% of the data ;Ensemble(loop)←− Ensemble trained with morphological features;Ensemble(loop+1)←− Ensemble trained with intensity features;Ensemble(loop+2)←− Ensemble trained with texture features;

CMat←− Compute confusion matrix for each of the classifier (to be used by the fusiontechnique);

OutputEnsemble(loop)←− classify validation set with Ensemble(loop);OutputEnsemble(loop+1)←− classify validation set with Ensemble(loop+2);OutputEnsemble(loop+2)←− classify validation set with Ensemble(loop+3);

OutputsS1←− Naive Bayes fusion off all 300 outputs;

%Stage 2;for loop:1:3:300 do

train←− Sample 80% of the patterns which are masses ;Ensemble(loop)←− Ensemble trained with morphological features;Ensemble(loop+1)←− Ensemble trained with intensity features;Ensemble(loop+2)←− Ensemble trained with texture features;

CMat←− Compute confusion matrix for each of the classifier (to be used by the fusiontechnique);

if pattern classified as mass in stage 1 thenOutputEnsemble(loop)←− classify validation set with Ensemble(loop);OutputEnsemble(loop+1)←− classify validation set with Ensemble(loop+1);OutputEnsemble(loop+2)←− classify validation set with Ensemble(loop+2);

OutputsS2←− Naive Bayes fusion off all three outputs;Final Output←− OutputsS2;

elseFinal Output←− OutputsS1;

Compute the accuracy for this fold using Final Output and Label.

Average the accuracy

59

Page 78: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

4.2.3 Fusion Techniques

In order to combine the outputs of the three classifiers which use different types of features (morpho-

logical, intensity and texture) and of the classifiers which use only a sample of the training data (instead

of the whole set), three combing techniques can be used: Majority Vote, Weighted Majority Vote and

Naive Bayes combination [38].

The first one, Majority Vote, as the name implies, is simply to pick the class which is output by the

majority of the classifiers. Given that there are binary classes in this work, let’s map the classes as

presented in equation 4.19 or 4.20.

ostage1 =

1, if mass

0, if healthy(4.19)

or

ostage2 =

1, if mass malignant

0, if mass benign(4.20)

Given this, in order to combine the outputs via a majority vote technique, one must simply average

them and choose the closest class. For example, let’s say that there are K classifiers, in this example,

k ∈ 1, 2, 3 and that o1 = 0, o2 = 1, o3 = 1. Then, by averaging, one obtains,

oexample =1

K

K∑k=1

ok = 0.666 (4.21)

Given that the value is bigger than 0.5, this pattern would be classified as a mass.

The second method, Weighted Majority Vote, uses the accuracy of each classifier and computes a

weight function based on it. This way, the classifiers have more power to make the final decision, if they

perform better. The weighted function is computed, as seen in equation 4.22, where, acck represents

the accuracy of each classifier k. For convenience, this function is normalized, as seen in equation 4.23.

b(k) = logacck

1− acck(4.22)

K∑k=1

b(k) = 1 (4.23)

Given binary classes, one must simply multiply the weight of each classifier by its output, as in

equation 4.24, for each pattern.

ocomb =

K∑k=1

bk × ok (4.24)

Using the same example as before, where o1 = 0, o2 = 1, o3 = 1, 4, let’s say that b1 = 0.58, b2 =

4in which the pattern is classified as a mass according to the Majority Vote technique

60

Page 79: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

0.21, b3 = 0.21, then, the combined output would be

ocomb = (0.58× 0) + (0.21× 1) + (0.21× 0) = 0.42 (4.25)

In which case, the pattern would be considered healthy tissue and not a mass.

The third combining method, Naive Bayes Combination, as the name implies, computes the combi-

natorial output using the Bayes rule. This method assumes independence between the classifiers and

its added value is that it computes the posterior probability of a pattern to be of a certain class, given the

output of the classifier. This probability is described in equation 4.26.

P (di = c|o) =P (di = c)P (o|di = c)

P (o)(4.26)

where P (di = c|o) is the probability of the correct label di to be of class c knowing the output vector

o, P (di = c) is the probability of the correct label di to be of class c, P (o|di = c) is the probability of

output vector o to happen, knowing that di = c and P (o) is the probability of the output vector o to

happen. Given the assumption of independence between the classifiers, the equivalence present in

equation 4.27 can be stated.

P (o|di = c) = P (o1, o2, ..., oK |di = c) =

K∏k=1

P (ok|di = c) (4.27)

Hence, one can say that,

P (di = c|o) =P (di = c)

∏Kk=1 P (ok|di = c)

P (o)(4.28)

and given that the denominator is independent from the class, one classifies a pattern as class c by

saying class c is the value that maximizes equation 4.29:

P (di = c|o) ∝

(P (di = c)

K∏k=1

P (ok|di = c)

)(4.29)

All the elements of equation 4.29 can be computed from the dataset or from the confusion matrix of

each classifier, by estimating the prior probability of class c (P (di = c)) as Nc/N , which is the number of

patterns belonging to class c over the number of total patterns and by computing the posterior probability

from the confusion matrix as described in equation 4.30, where the (c, s)th entry of matrix cm is the

number of elements in the dataset whose true class is c and are classified as s.

P (ok|di = c) =cmk(c, s)

Nc(4.30)

61

Page 80: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

The final expression to compute the combinatorial label can be found in equation 4.31 and 4.32.

P (di = c|o) ∝NcN

∏Kk=1 cmk(c, s)

NKc

∝1

NK−1c

K∏k=1

cmk(c, s) (4.31)

ocomb = argmax (P (di = c|o))c

(4.32)

62

Page 81: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

5Experimental Results

Contents

5.1 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

5.2 Pre-Processing and Segmentation Results . . . . . . . . . . . . . . . . . . . . . . . . 66

5.3 Classification Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69

5.4 Comparison of Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73

63

Page 82: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

64

Page 83: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

5.1 Data

The database used for this work was compiled by South Florida University and is called Digital

Database for Screening Mammography(DDSM) [43] [44]. It is composed of 1950 lesion-free mammo-

grams, 836 mammograms containing malignant masses and 399 containing benign ones, as displayed

in figure 5.1. In DDSM, the lesions are previously labelled by physicians and confirmed via a biopsy. A

certain error in this database has to be assumed, for although each exam might have been reviewed

by other experts, the sensitivity of a doctor is lower than 93% [45], when locating lesions, in this type of

exam. This means that there might be masses that are present in the images but that are not accounted

for. There are also several exams with microcalcification clusters, which were not used, for these type of

lesions fall outside of the scope of this thesis.

Figure 5.1: Histogram of the 3 classes of exams contained in DDSM database.

One obstacle when using this database is the fact that the files are compressed in ljpeg format

(lossless jpeg). This format is not widely known, therefore, there is a need to convert them to a format

readable by MATLAB2016a (or other applications). Along with the database, a C source code is available

to convert the images from .ljpeg to .pgm. This last format is not commonly recognized by a standard

image viewer, but it can be written to other formats in order to be viewed. In this thesis, a virtual machine

with Ubunto as an operating system is used to run the C code to convert the files to .pgm. The total

size of files is around 220Gb, so a total of 1 day of computation time is required to convert the entire

65

Page 84: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

database.

The contents of DDSM database are originated from two locations: Massachusetts General Hospital

(MGH) and Wake Forest University School of Medicine (WFUSM). Although the race of each patient is

not specified per exam, table 5.1 presents the race breakdown of the female patient population of both

medical facilities.

Table 5.1: The female population at MGH and WFUSM by race

Race MGH (%) WFUSM (%)Asian 2.06 0.20Black 4.12 20.40Spanish (origin) 6.55 1.80American Indian 0.00 0.10Other 0.75 0.10Unknown 30.34 0.30White 56.18 77.00

So far, no study has found a direct connection between breast cancer and race, but it is important to

be aware that the exams presented in this database are, in majority, from white women.

5.2 Pre-Processing and Segmentation Results

This section describes the results of the bilateral filter and segmentation method presented in section

3.2. The bilateral filter built in this work is custom built, according to the literature, as described in section

3.1. The filter is used with a window of analysis of 81 × 81 pixels, a standard deviation of 20 pixels (for

the spatial distance filter) and a standard deviation of 12% of the maximum value of the image (for the

intensity based filter).

The effects of this filter can be seen in figures 5.2(a) to 5.4(b), where its influence over the segmen-

tation stage is exemplified. The effects of the bilateral filter can be divided into three categories: the filter

improves the final result, the filter has no influence and the use of the filter outputs a worse result. The

segmentation algorithm (SBF) is provided by the writers of literature [33] and adapted to this experiment.

For these experiments, the segmentation algorithm used a total of 32 orientation points.

The best case scenario, where the bilateral filter reduces the number of false positives, is presented

in figures 5.2(a) and 5.2(b). The figures exemplify a case of a patient’s breast with cancer (seen in blue).

The segmentation algorithm starts by considering 9 regions to be ROI, none of which is the correct

lesion, whereas if the image is filtered, only 3 ROIs are considered (including the correct one). This is a

proof that the use of the bilateral filter reduces the number of false positives.

Other cases occur as well, where the bilateral filter does not influence the final result, by not reducing

the number of ROIs, as is presented in figures 5.3(a) and 5.3(b).

66

Page 85: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Lastly, the final case that occurs is where the application of the bilateral filter has a negative impact

and the real lesion is not detected as a ROI. This implies that the use of the bilateral filter, in some

cases, reduces the accuracy of the whole system. Such a case is presented in figures 5.4(a) and 5.4(b).

(a) Unfiltered image (b) Filtered image

Figure 5.2: Effect of filtering on the segmentation of a breast image with a malignant lesion a) without the bilateralfilter b) with the bilateral filter. In this figure, it can be seen the output of the segmentation in green andthe ground truth in blue.

67

Page 86: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

(a) Unfiltered image (b) Filtered image

Figure 5.3: Effect of filtering on the segmentation of a breast image with a malignant lesion a) without the bilateralfilter b) with the bilateral filter. In this case, the filter had no effect on the segmentation output.

(a) Unfiltered image (b) Filtered image

Figure 5.4: Effect of filtering on the segmentation of a breast image with a malignant lesion a) without the bilateralfilter b) with the bilateral filter. In this case, the segmentation algorithm does not locate the correct ROIin the filtered image.

This filter influences the final FP value, reducing the number of ROIs found in 20%. The system can

be parametrized for a smaller FP with the trade-off of reducing the TP rate, for some lesions would not

be detected in the segmentation stage.

68

Page 87: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

After the segmentation algorithm is applied, the features mentioned in section 4.1 are extracted in

order to build a feature dataset, with three possible labels: healthy tissue, benign mass and malignant

mass.

5.3 Classification Results

5.3.1 Single Model Decision Tree Experiment

As mentioned in section 4.2.1, a first experiment is done using only one DT. This experiment was

done to be compared with the others, for given that the AdaBoost classifier is built with single-node deci-

sion trees, a single decision tree is the simplest form of this type of classifier. This tree is computed with

MATLAB2016a’s built-in function and pruned as much as possible, without increasing the classification

error, using 10-fold cross-validation. Pruning is used in order to avoid over-fitting the tree, by growing

it too large. Figures 5.5(a) and 5.5(b) exemplify the same tree (not pruned and pruned, respectively),

where it is possible to see that the tree’s size is reduced. After training and cross-validating the tree, the

confusion matrix present in table 5.2 represents the output of this classifier, which obtained an average1

accuracy of 82%.

(a) Decision tree (b) Pruned Decision Tree

Figure 5.5: DT a) not pruned, b) pruned, with three possible outputs: 0,1 and 2 or healthy tissue, malignant massand benign mass.

1Average of the 10 cross-validation folds

69

Page 88: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Table 5.2: Confusion matrix of the decision tree presented in the Single Model Decision Tree experiment. Thismodel obtained a classification rate of 82%.

Output Labels

Healthy Tissue Benign Mass Malignant Mass

Target Class

Healthy Tissue 1760 83 107

Benign Mass 136 189 125

Malignant Mass 134 56 595

DDSM is not a balanced database, as seen in section 5.1. This means that some classes of patterns

are more present in the dataset than others, which will affect the results when using a DT. This type of

classifier has an unstable behaviour when used with unbalanced datasets, which privileges the patterns

which are most common (and the reverse for the uncommon ones).

In tables 5.2 and 5.3 can be observed that this model performs better when classifying healthy tissues

and performs significantly worse when classifying benign masses, which are the most common class

of patterns and the most rare, respectively. This is the result of the unstable behaviour from the DT

mentioned earlier.

Table 5.3: Table presenting the per class accuracy results of the Single Model Decision Tree experiment.

Accuracy of the Single Model Decision Tree

General 0.82

Malignant 0.80

Benign 0.47

Healthy Tissue 0.91

5.3.2 Decision Tree Ensemble Experiment

In this experiment, a two stage hypothesis was tested as presented in section 4.2.1. It is composed

of 6 DTs, grown with a stopping criteria of 16 splits. In each tree, only a specific set of features is used

as training set: morphological, intensity based or texture ones. This experiment is tested using 10 fold

cross-validation technique and the trees are also pruned. The confusion matrix for this experiment can

be seen in table 5.4 and its accuracy is 82%.

A first analysis to be done is that this schema performs (in general) the same way as the single

decision tree presented in section 5.3.1. An interesting aspect that can be observed through tables 5.4

and 5.5 is the influence of the fusion technique. Note that, for both types of mass and all classes put

together, the output results is always improved when using the fusion of the 3 model than when using

each model in separate. With it, it can also be observed that the effect of the data being unbalanced

70

Page 89: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

is attenuated with the fusion technique, increasing the per class accuracy for the less common class

(benign mass) and decreasing for other two. An analysis on the relationship between each model can

also be retrieved, where it can be noted that the morphological features model performs best (compared

to the other models). Table 5.5 displays these relationships.

Table 5.4: Confusion matrix of the classifier presented in the Decision Tree Ensemble Experiment. The final classi-fier obtained a classification rate of 82%.

Output Labels

Healthy Tissue Benign Mass Malignant Mass

Target ClassHealthy Tissue 1763 99 88

Benign Mass 99 255 96

Malignant Mass 105 80 600

Table 5.5: Table presenting the accuracies per class and per model obtained from the results of the Decision TreeEnsemble Experiment.

Morphological Feat. Intensity Feat Texture Feat Classifier ensemble

General 0.80 0.73 0.71 0.82

Malignant 0.62 0.42 0.43 0.77

Benign 0.41 0.45 0.18 0.57

Healthy Tissue 0.97 0.92 0.94 0.90

Table 5.6: Confusion matrix of the morphological, intensity and texture features model presented in the DecisionTree Ensemble Experiment. Each model obtained a classification rate of 80%, 73% and 71%, respec-tively.

Output Labels

Healthy Tissue Benign Mass Malignant Mass

Target Class - Morphological featuresHealthy Tissue 1879 43 28

Benign Mass 193 183 74

Malignant Mass 211 87 487

Target Class - Intensity featuresHealthy Tissue 1801 112 37

Benign Mass 193 202 55

Malignant Mass 249 207 329

Target Class- Texture featuresHealthy Tissue 1834 75 41

Benign Mass 292 84 74

Malignant Mass 331 122 332

In order to comprehend the relationship between each model and each class, one can observe the

71

Page 90: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

confusion matrix of each of them, presented in table 5.6. In this matrix, the effect of the unbalanced

data can be observed by the amount of patterns miss classified as healthy tissue. In the limit, the

morphological features and the texture features models classify patterns corresponding to the benign

classes more often as healthy tissues, than as benign masses themselves.

5.3.3 AdaBoost Ensemble Experiment

The third and last experiment is an optimized/modified version of the second one. In this case,

AdaBoost replaces the DTs in the classifier ensemble. 600 AdaBoost were trained, in which each had

a maximum of 10 learners. The input of each classifier was 80% of the total training set, obtained using

bootstrap sampling with replacement. Table 5.7 presents the confusion matrix of the final classifier. With

it, significant improvements can be spotted, when comparing this AdaBoost Ensemble experiment with

Decision Tree Ensemble experiment: the correct labels for benign masses and healthy tissue increased.

Also, these settings reduce, in a great deal, the number of patterns miss classified as mass (both benign

and malignant).

Table 5.7: Confusion matrix of the AdaBoost Ensemble Experiment classifier. This model obtained a classificationrate of 85%.

Output LabelsHealthy Tissue Benign Mass Malignant Mass

Target ClassHealthy Tissue 1863 59 28Benign Mass 113 269 68

Malignant Mass 81 153 551

Table 5.8 allows us to analyse the relationship between models and classes. It can be observed that

using AdaBoost improves the results of the final classifier for all 3 features models, and, consequently,

for the general accuracy as well. When using AdaBoost, the fusion technique does not always increase

the result of the classifier ensemble, when comparing to each model in separate, for the effect of the

unbalanced data is contradicted in the AdaBoost itself, in the weight pattern attribution (see section

4.2.2). Given this, the fusion technique outputs a better result than any of the models, when the per

class accuracy is low and might not increase the general accuracy if the per class accuracy is high.

72

Page 91: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Table 5.8: Table presenting the accuracies per class and per model obtained from the results of the AdaBoostEnsemble Experiment classifier.

Morphological Feat. Intensity Feat Texture Feat Classifier ensemble

General 0.84 0.76 0.75 0.85

Malignant 0.74 0.59 0.62 0.70

Benign 0.46 0.27 0.20 0.60

Healthy Tissue 0.97 0.94 0.92 0.96

As explain in section 4.2.2, each AdaBoost contains 10 learners, each with a corresponding weight

that is posteriorly used when computing the label of the AdaBoost. This type of classifier works in a

coarse-to-fine manner, so, for the first learner, the classification weights associated to the patterns is

uniform, which means that the first learners will have the highest weights associate to them, as seen in

figure 5.6(a). Equally, the same happens with the learners, broken down by model, as seen in figure

5.6(b). In section 4.2.2 it is proven that the weights of each leaner are proportional to the learner’s

classification rate and such can be observed by the information contained in table 5.8 and figure 5.6(b),

where it can be seen that the model which uses morphological features has both the highest weights

and highest accuracy.

(a) Total average of the weights of the learners (b) Average of the weights of the learners by model

Figure 5.6: Relative weights of the learners of the AdaBoost classifier a) in total b) by model.

5.4 Comparison of Results

In order to compare the results of the three experiments, the results can be seen from two different

perspectives:

73

Page 92: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

• A mathematical perspective, where we compare the classification results for the three classes,

independently of what they mean, as reported in section 5.3.

• A commercial/medical perspective, where we analyse and compare only the accuracy of the

classifier in distinguishing healthy tissue from masses (output of stage 1). Given this, it is consid-

ered irrelevant the distinction between benign mass and malignant mass, since both information

should be reported to the doctor.

The first analysis to be done is to compare the three experiments as presented in section 5.3. This

way it is possible to identify the settings which would output the best results, from a mathematical point

of view. Table 5.9 contains the general and per class accuracies of the three experiments. Here, one

can observe that the settings from AdaBoost Ensemble experiment output the best results in general,

when classifying benign lesions or healthy tissue, behaving the worse for malignant lesions.

Table 5.9: Table for comparison of general accuracies and per class accuracy of the three experiments

Single Model Decision Tree Decision Tree Ensemble AdaBoost EnsembleACC General 0.82 0.82 0.85

ACC Malignant 0.80 0.77 0.70ACC Benign 0.47 0.57 0.60

ACC Healthy Tissue 0.91 0.90 0.96

A second analysis is done, considering the commercial/medical perspective. Let’s look at the fact that

this CAD system deals with a diagnosis problem, where it can be said that if the CAD miss classifies

a benign mass as a malignant one, it is not the same as mistaking a malignant mass for a healthy

tissue. With this assumption, it is possible to adapt the classifier to reality and real life environments

by considering only two classes (healthy tissue or mass), in order for the output of the classifier to be

studied under a different perspective.

Table 5.10 presents the confusion matrix off all three experiments, where the patterns labelled as

”benign mass” or ”malignant mass” were simply labelled as ”mass”. The data provided by this table sup-

ports the fact that the settings of the AdaBoost Ensemble experiment perform better when appropriating

the CAD system to real life, considering only two classes. Although these settings performed worse for

malignant class, when joining both lesions as one single class, the AdaBoost based classifier ensemble

outperforms the other two, achieving higher classification rates for both classes.

74

Page 93: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Table 5.10: Confusion matrix for all three experiments, considering only two masses.

Output LabelsHealthy Tissue Mass

Target Class - Single Mode Decision Tree Healthy Tissue 1745 205TP = 0.79 — FP = 0.11 — ACC = 0.85 Mass 266 969

Target Class - Decision Tree Ensemble Healthy Tissue 1763 187TP = 0.84 — FP = 0.09 — ACC = 0.88 Mass 204 1031

Target Class - AdaBoost Ensemble Healthy Tissue 1863 87TP = 0.84 — FP = 0.04 — ACC = 0.91 Mass 194 1041

75

Page 94: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

76

Page 95: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

6Conclusions and Future Work

77

Page 96: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

78

Page 97: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

This thesis presents a Computer Aided Diagnosis (CAD) system that identifies lesions in mammo-

graphies. The system is built to filter and segment images in order to find regions of interest (ROIs) to

later classify them into healthy tissue, benign mass or malignant mass.

This work uses several steps to locate and identify each lesion. It uses a bilateral filter to pre-process

the image, which enhances the edges of objects within it. This filter can be computationally heavy, for it

computes both the spatial distance between the pixels and their intensity difference. The bilateral filter

uses these two measurements and smooths the inside and the outside of the objects, removing noise

and irrelevant information, without losing the edges.

After preprocessing, probable regions of being a lesion (ROIs) are identified and segmented. This is

based on the gradients of the image, for the ROIs are located by finding the local maxima of convergence

of such gradients. In this work, Sliding Band Filter (SBF) is used for this step. This algorithm takes into

account a radius range, in order to reduce the possibilities of the search and decreases, in a great

deal, the number of false positives, for it provides the possibility to consider only certain regions as

ROIs, instead of assuming every possible location. By using the bilateral filter, prior to this segmentation

algorithm, the number of false positives is reduced in 20%.

After the ROIs have been segmented, different information about each ROI can be extracted. In this

work, morphological, intensity-based and texture information is analysed.

This information is used to classify each ROI into healthy tissue, benign or malignant mass. In this

work, different classifier ensembles have been tested to classify such ROIs. The first experiment is a

Single Model Decision Tree, with all features available. This set-up outputs a maximum accuracy of

82% and, in these conditions, it has a higher per class accuracy for healthy tissues. The classifier

ensemble of the second experiment, Decision Tree Ensemble, uses a 2-stage schema and combines 3

classifiers per stage, where the first classifies the ROIs into healthy tissue or lesion and the second, in

benign mass or malignant one. This experiment results in a classifier with an accuracy of 82%, which

is similar to the Single Model Decision Tree. This points to the conclusion that this model might be

using the ”weaknesses” of its settings without using the ”strengths”: it divides the dataset by features

(which reduces the number of features available for the classifier) (weakness) but it combines only three

models, which it not enough to leverage on the fusion technique used to combine the outputs of each of

them (strength).

The third one also uses a 2-stage classifier, but this time, it uses 300 classifiers, instead of just

3, where in each is input an 80% sample of the data, instead of the whole set. This means that not

all classifiers will output the same results. With these settings, a general classification rate of 85% is

achieved. As shown in section 5.4, when considering benign masses and malignant masses to be one

class, it is also the settings of the AdaBoost Ensemble experiment the ones which output the best results.

79

Page 98: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

As seen, when using the CAD present in this thesis, the results of a diagnosis would improve, in

general. Nevertheless, many other hypothesis can be studied and some modifications can be done to

this work that might improve its performance. One analysis to further be done is the influence of the

parameters of the initial stages in the final classification. The minimum and maximum radius of the

lesions considered in the bilateral filter and segmentation algorithm or the thresholds of this last one,

were chosen based on their visual results and in the amount of ROIs found and their location. This can

be perfected if a system is built to analyse the direct influence of these parameters (at the beginning of

the stream) in the output of the final classifier. Another subject where this work can be improved is in

the analysis of the lesions themselves. Given that it falls out its scope, this thesis leaves much room

for exploring the different kinds of malignant and benign masses and for introducing expert medical

knowledge about their physical aspects or typical locations in order to optimize the segmentation and/or

classification algorithms. An extended possibility of this aspect is to introduce several other classes,

where instead of simply malignant or benign lesions, the system can be searching for micro invasive1,

ductal invasive1, lobular invasive1 lesions, etc. This creates the possibility to optimize the procedures to

find each particular class, according to its characteristics.

The study for using the CAD for its final purpose, where it is more significant to detect the malignancy

of a lesion, than to exclude healthy tissue, can be further continued by exploring the cost of classification

between classes of the classifiers. As mentioned in section 5.4, the best settings presented in this work

achieved TP and FP rate values of 84% and 4%, respectively. This FP value is low enough for a further

research to increment the TP value (and consequently, increment the FP value), in order to explorer the

best settings, not just in terms of accuracy, but in terms of appropriating the system to its final use.

All the suggestions made above, aim to bring the work on CAD systems closer to reality and real life

environments.

1Examples of different types of malignant tumours

80

Page 99: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

Bibliography

[1] B. Stewart, C. P. Wild et al., “World cancer report 2014,” World, 2015.

[2] W. C. R. F. International. Worldwide data. [Online]. Available: http://www.wcrf.org/int/

cancer-facts-figures/worldwide-data

[3] C. R. UK. Breast cancer statistics. [Online]. Available: http://www.cancerresearchuk.org/

health-professional/cancer-statistics/statistics-by-cancer-type/breast-cancer

[4] N. C. Institute. Female breast cancer. [Online]. Available: http://seer.cancer.gov/statfacts/html/

breast.html

[5] P. H. England, “Cancer survival in England by stage,” Jul 2014, pp. 8–14.

[6] S. Vedantham, L. Shi, A. Karellas, K. E. Michaelsen, V. Krishnaswamy, B. W. Pogue, and K. D.

Paulsen, “Semi-automated segmentation and classification of digital breast tomosynthesis recon-

structed images,” in 2011 Annual International Conference of the IEEE Engineering in Medicine

and Biology Society, Aug 2011, pp. 6188–6191.

[7] B. D. Fornage, “Interventional ultrasound of the breast,” in Breast Ultrasound Update. Karger

Publishers, 1994, pp. 260–264.

[8] B. Caroline and N. Vaijayanthi, “Computer aided detection of masses in digital breast tomosyn-

thesis: A review,” in Emerging Trends in Science, Engineering and Technology (INCOSET), 2012

International Conference on, Dec 2012, pp. 186–191.

[9] Radiopedia.org. Mammography views. [Online]. Available: http://radiopaedia.org/articles/

mammography-views

[10] Canada.com. Category 3 breast mammography. [Online]. Available: http://o.canada.com/health-2/

when-mammograms-lie-how-cancer-can-sneak-past-our-greatest-screening-techniques

[11] C. J. Rose, “Statistical models of mammographic texture and appearance,” Ph.D. dissertation, Cite-

seer, 2005.

81

Page 100: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

[12] C. C. Society. Mammography. [Online]. Available: http://convio.cancer.ca/site/PageServer?

pagename=SSL ON L BreastCancer ScreeningTest#.VyNXqDArJIJ

[13] J. Peixoto, E. Canella, and A. Azevedo, “Mamografia: Da pratica ao controle,” Rio de Janeiro: INCA,

2007.

[14] A. Colaco, “Cancro da mama - diagnosticos e desafios,” Faculdade de Ciencias da Universidade

de Lisboa, Lisbon, Portugal, 2016.

[15] N. Matela, “Tomossıntese no diagnostico do cancro da mama,” Instituto de Biofısica e Engenharia

Biomedica - Faculdade de Ciencias da Universidade de Lisboa, Lisbon, Portugal, 2014.

[16] G. Global Business Unit, Siemens AG Medical Solutions, “Data sheet mammomat inspiration,” pp.

10–11, 2015.

[17] I. Duarte, F. Janela, L. Caldeira, F. Soares, and J. Silva, “3D medical image reconstruction on digital

breast tomosynthesis,” in Bioengineering (ENBENG), 2012 IEEE 2nd Portuguese Meeting in, Feb

2012, pp. 1–6.

[18] W. R. Associates. New low dose 3D mammography. [Online]. Available: http://www.

washingtonradiology.com/patient-education/mammo-lowdose.asp

[19] J. Wei, B. Sahiner, L. M. Hadjiiski, H.-P. Chan, N. Petrick, M. A. Helvie, M. A. Roubidoux, J. Ge, and

C. Zhou, “Computer-aided detection of breast masses on full field digital mammograms,” Medical

physics, vol. 32, no. 9, pp. 2827–2838, 2005.

[20] R. Alam, F. Younesi, and M. R. Alam, “Computer-aided mass detection on digitized mammograms

using a novel hybrid segmentation system,” Int J Biol Biomed Eng, vol. 3, no. 4, pp. 51–58, 2009.

[21] H.-P. Chan, J. Wei, Y. Zhang, M. A. Helvie, R. H. Moore, B. Sahiner, L. Hadjiiski, and D. B. Kopans,

“Computer-aided detection of masses in digital tomosynthesis mammography: comparison of three

approaches,” Medical physics, vol. 35, no. 9, pp. 4087–4095, 2008.

[22] Bilateral filtering for grey and color images. [Online]. Available: http://homepages.inf.ed.ac.uk/rbf/

CVonline/LOCAL COPIES/MANDUCHI1/Bilateral Filtering.html#Index

[23] R. Thyagarajan and S. Murugavalli, “Segmentation of digital breast tomograms using clustering

techniques,” in 2012 Annual IEEE India Conference (INDICON), Dec 2012, pp. 1090–1094.

[24] X. Qin, G. Lu, I. Sechopoulos, and B. Fei, “Breast tissue classification in digital tomosynthesis

images based on global gradient minimization and texture features,” in SPIE Medical Imaging.

International Society for Optics and Photonics, 2014, pp. 90 341V–90 341V.

82

Page 101: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

[25] I. Reiser, R. Nishikawa, A. Edwards, D. Kopans, R. Schmidt, J. Papaioannou, and R. Moore, “Auto-

mated detection of microcalcification clusters for digital breast tomosynthesis using projection data

only: a preliminary study,” Medical physics, vol. 35, no. 4, pp. 1486–1493, 2008.

[26] M. R. Hejazi and Y.-S. Ho, “Automated detection of tumors in mammograms using two segments

for classification,” in Pacific-Rim Conference on Multimedia. Springer, 2005, pp. 910–921.

[27] H. Kobatake and S. Hashimoto, “Convergence index filter for vector fields,” IEEE Transactions on

Image Processing, vol. 8, no. 8, pp. 1029–1038, Aug 1999.

[28] D. Saraswathi and E. Srinivasan, “A high-sensitivity computer-aided system for detecting micro-

calcifications in digital mammograms using curvelet fractal texture features,” Computer Methods in

Biomechanics and Biomedical Engineering: Imaging & Visualization, pp. 1–11.

[29] D. Saraswathi and Srinivasan, “A CAD system to analyse mammogram images using fully complex-

valued relaxation neural network ensembled classifier,” Journal of medical engineering & technol-

ogy, 2014.

[30] W. K. Pratt, “Digital image processing, new-york,” NY: John Wiley and Sons, 1991.

[31] D. H. Kim, S. T. Kim, W. J. Baddar, and Y. M. Ro, “Feature extraction from bilateral dissimilar-

ity in digital breast tomosynthesis reconstructed volume,” in Image Processing (ICIP), 2015 IEEE

International Conference on, Sept 2015, pp. 4521–4524.

[32] N. Petrick, H.-P. Chan, D. Wei, B. Sahiner, M. A. Helvie, and D. D. Adler, “Automated detection of

breast masses on mammograms using adaptive contrast enhancement and texture classification,”

Medical Physics, vol. 23, no. 10, pp. 1685–1696, 1996.

[33] T. Esteves, P. Quelhas, A. M. Mendonca, and A. Campilho, “Gradient convergence filters and a

phase congruency approach for in vivo cell nuclei detection,” Machine Vision and Applications,

vol. 23, no. 4, pp. 623–638, 2012.

[34] L. Jacques, “Image segmentation and classification,” Ecole Polytechnique de Louvain - Universite

Catholique de Louvain.

[35] Mathworks. Intermediate differences gradient. [Online]. Available: http://www.mathworks.com/help/

images/ref/imgradient.html?s tid=srchtitle

[36] Q. Zheng, “Mathematical morphology - set-theoretic representation for binary shapes,” Language

and Media Processing Lab - Center for Automation Research, University of Maryland College Park.

[37] M. Davis and F. Sahin, “Hog feature human detection system,” in Systems, Man, and Cybernetics

(SMC), 2016 IEEE International Conference on. IEEE, 2016, pp. 002 878–002 883.

83

Page 102: COMPUTER AIDED DIAGNOSIS SYSTEM IN DIGITAL …

[38] L. I. Kuncheva, Combining pattern classifiers: methods and algorithms. John Wiley & Sons, 2004.

[39] P. Dupont, “Decision tree learning,” ICTEAM - Universite Catholique de Louvain, Louvain-la-neuve,

Belgium.

[40] P. Viola and M. J. Jones, “Robust real-time face detection,” International journal of computer vision,

vol. 57, no. 2, pp. 137–154, 2004.

[41] B. Kegl, “The return of adaboost. mh: multi-class hamming trees,” arXiv preprint arXiv:1312.6086,

2013.

[42] L. Breiman, “Bagging predictors,” Machine learning, vol. 24, no. 2, pp. 123–140, 1996.

[43] D. K. R. M. Michael Heath, Kevin Bowyer and W. P. Kegelmeyer, “The digital database for screening

mammography.” Medical Physics Publishing, 2001, pp. 212–218.

[44] D. K. W. P. K. R. M. K. C. Michael Heath, Kevin Bowyer and S. MunishKumaran, “Current status

of the digital database for screening mammography.” Kluwer Academic Publishers, 1998, pp.

457–460.

[45] S. Hofvind, B. Geller, J. Skelly, and P. Vacek, “Sensitivity and specificity of mammographic screening

as practised in vermont and norway,” The British journal of radiology, 2014.

84