Conference PaperPDF Available

Breast Cancer Classification and Proof of Key Artificial Neural Network Terminologies

Authors:

Abstract and Figures

Classification is one of the interesting areas in the academic field of Neural Networks. Artificial Neural Networks (ANNs) have been extensively used in pattern recognition and classification of data in the supervised and unsupervised environment. The ANNs use advanced concepts of computer science where a machine mimics human intelligence while learning from possible experience. To make a machine self-adaptive and autonomous, the machine is properly trained on a training data-set and then subsequently tested on new data. The excellent quality of training of ANNs typically depends on the underlying architecture of the network they employ, for a specific instance, a considerable number of deep layers, number of key nodes in each distinct layer, epoch size, and activation function. In this academic paper, the practical importance of these architectural components is carefully investigated. This paper is precisely about providing a solution that how ANNs can help us in Breast Cancer Classification. Furthermore, sufficient proofs of some extremely important terminologies used in ANNs are also discussed which will clarify the important concepts of ANNs.
Content may be subject to copyright.
2019 13th International Conference on Mathematics, Actuarial Science, Computer Science and Statistics (MACS)
Breast Cancer Classification and Proof of Key
Artificial Neural Network Terminologies
Nisar Ali, Shahab Ansari, Zahid Halim, Raja Hashim Ali, Muhammad Faizan Khan, Mohsin Khan
Faculty of Computer Science and Engineering
Ghulam Ishaq Khan Institute of Engineering Sciences and Technology
{nisarali4465, ansarishahab111, halimzahid, raja.hashim.ali1, faizan4465.khan, mykhan911}@gmail.com
Abstract—Classification is one of the interesting areas in the
academic field of Neural Networks. Artificial Neural Networks
(ANNs) have been extensively used in pattern recognition
and classification of data in the supervised and unsupervised
environment. The ANNs use advanced concepts of computer
science where a machine mimics human intelligence while
learning from possible experience. To make a machine self-
adaptive and autonomous, the machine is properly trained on a
training data-set and then subsequently tested on new data. The
excellent quality of training of ANNs typically depends on the
underlying architecture of the network they employ, for a specific
instance, a considerable number of deep layers, number of key
nodes in each distinct layer, epoch size, and activation function.
In this academic paper, the practical importance of these
architectural components is carefully investigated. This paper is
precisely about providing a solution that how ANNs can help us
in Breast Cancer Classification. Furthermore, sufficient proofs
of some extremely important terminologies used in ANNs are
also discussed which will clarify the important concepts of ANNs.
Index Terms—Artificial Neural Network, Key Terminologies,
Breast Cancer Classification, and Proofs
I. INTRODUCTION
Breast cancer is a type of disease where malignant (cancer)
cells form in the tissues of the breast. It starts when breast cells
begin to grow out of control. A tumor appears due to these
cells which can easily be seen on an x-ray. Breast cancer exists
almost entirely in women, but there is a chance that men can
get breast cancer, too [1].
Among women, the most commonly diagnosed cancer is
of the breast [2]. In the United States, one in eight women
is diagnosed with this disease in her lifetime. This disease
remains the second leading cause of cancer deaths in women.
Each year over 252,500 women in the United States are
diagnosed with breast cancer and more than 40,000 are died
because of this disease. This disease is rare in men, but it is
estimated that over 2,500 men are diagnosed with this disease
and 500 die every year [3]. On average, every two minutes a
woman is diagnosed with this fatal disease and a woman has
died every 14 minutes. Over 3.3 million survivors are alive in
the United States with breast cancer [4].
A lot of work has been done to stop this disease from
causing harm [5] [6] [7] [8]. It takes time to diagnose breast
cancer due to which it spreads and the chances of death
increase. It is very important to diagnose such a fatal disease
quickly and for this, there should be a computerized system
that will help to diagnose breast cancer faster [9]. For this
many techniques have been proposed and one of this is
Artificial Neural Networks. This method has been used widely
to classify breast cancer to help the community.
The rest of the paper is structured as follows. In section II
we introduce the Artificial Neural Network. In section III
related work is discussed. Our methodology is proposed in
section IV. Results and proofs are reported in section V.
Finally, conclusions are drawn in section VI.
II. ART IFI CI AL NEURAL NE TWORK
Artificial neural network has provided an exciting alternative
method for solving a variety of problems in different fields
of science and engineering. An artificial neural network is
defined as the interconnection pattern between the different
layers of neurons. Each connection between neurons has a
weight [8]. Each neuron is represented by a node and these
nodes are connected through edges which makes the output
of one neuron, the input of others. Two main types of neural
networks are given below:
Feed-Forward Network
Feedback network
A. Feed-Forward Network
Feed-forward network [10] consists of a perceptron which
are organized in the form of layers. A feed-forward network
is a simple type of neural network which consists of an input
layer, an output layer, and one or more hidden layers. Each
perceptron in a single layer has no connection with each
other. Furthermore, each perceptron in a layer is connected
to all perceptrons on the next layer. The main advantage of
this network is that it learns to evaluate and recognize input
patterns [11]. In this network, the information moves only in
one direction i.e. forward, from the input nodes data goes
through hidden nodes (if any) and then to the output nodes.
Fig 1 shows the basic model of a feed-forward network.
B. Feedback Network
In this type of network [12], the output goes back into the
network to achieve the best result. Feedback networks are dy-
namic. Signals are traveling in both directions by introducing
loops in this network. The network remains at the equilibrium
point until the input changes. Feedback networks are used by
the internal systems for the correctness of errors [13]. Fig 2
shows the basic model of the feedback network.
978-1-7281-4956-1/19/$31.00 c
2019 IEEE
C. Artificial Neural Network Training
An artificial neural network is a computing system inspired
by the human brain. A human brain learns from its surround-
ings and make decisions using knowledge. An artificial neural
network can be developed just like a human brain by training
it with proper data-set so that it can be able to solve unseen
problems related to the specific data-set. A neural network
can be trained by different methods. A feed-forward network
can be trained by using Conjugate Gradient [14] and Back
Propagation (BP) method [15], the most common learning
algorithms. On the other hand, the feedback network can be
trained by using real-time recurrent learning algorithms [16]
[17].
Fig. 1. Basic Model of Feed Forward Network
Fig. 2. Basic Model of Feedback Network
III. REL ATED W OR K
In the last decade, a lot of work has been done for the
detection of breast cancer. Today, several screening techniques
are used to detect breast cancer like positron emission tomog-
raphy (PET) [18], CT scan [19], X-Ray [20], ultrasound [21],
mammogram [22] etc. These techniques have their advantages.
Mammogram is the most reliable and popular technique. But,
this technique has some serious limitations. About 30 of
the breast grazes couldnt be spotted in mammograms during
screening. This thing led the researchers towards developing
an automated computational system for breast cancer diagnosis
[23].
Artificial neural networks [24] are used for the diagnosis
of breast cancer since the past few years and their accuracy
rate is very significant as compared to other non-computational
diagnostics. For classification Support Vector Machine (SVM)
[25], Self-Organizing Map (SOM) [26], Probabilistic Neural
Network [27], General Regression Neural Network [28], Ra-
dial Basis Network (RBN) [29] and Multi-Layer Perception
(MLP) [5] have also been applied on the said task. The
obtained results show that General Regression Neural Network
has been the most accurate rate in identifying the nature
of the input as compared to others. Its accuracy rate was
98.8% respectively [30]. Xin Yao carried out a negative
correlation algorithm [31] which was able to automatically
decompose and solve the problem. He prominently mentioned
two unique approaches like an evolutionary approach that was
able to automatically design a compact neural network. The
second unique approach was the ensemble which was precisely
in progress but was carefully designed to properly address
massive problems [32].
IV. METHODOLOGY
Wisconsin data set [33] is used in the training and testing
of Neural Networks. In this data set first column accurately
indicates the id of each row, the next nine columns indicate
authentic data and the last column accurately represents the
desired output against each row. Each specific column among
the nine columns uniquely represents a specific property of a
cell i-e Clum Thickness, Bare Nuclei, Normal Nucleoli, etc.
The output column indicates whether it is malignant, means
4 or benign, means 2. We need separating ids, authentic data,
which represent properties of a cell and output data from given
data set before we train Neural Network. There are precisely
some unknown values in the given data set which are gently
replaced by 5 in the following experiments.
Significant key parameters which are used to set architecture
in training of Neural Networks are:
Activation Functions
Hidden Layers
Learning Functions
Mean Square Error
Epochs
V. RE SU LTS AND ANA LYSIS
The desired results of each experiment are analyzed by
checking the performance of the neural networks through
accuracy, precision, and efficiency along with proofs.
A. First Hypothesis
Training of Neural Network on half 50% of provided
data, in which 32.5% of selected data indicates Benign and
17.5% shows Malignant, would be much more effective as
compare favorably to randomly chosen data from given data
set. Successful experiments which have been precisely carry
out are adequately explained in Table I. The key parameters
which are constant for the experiment are MSE =0.01 and
Epochs =100
An immense difference in the performance of Neural Net-
works, if we compare accuracy favorably of experiments
being done under given hypothesis and randomly chosen data.
Nearly all of the successful experiments gave more than
96% of remarkable precision. The appropriate environmental
setup for the training of Neural Network, for the above-given
hypothesis, has been communicated below.
Hidden Layers: According to the accuracy rate of the
above experiments, it has been carefully observed that
the specific number of deep layers may typically vary
between 1 and 3 and the considerable size of deep layers
can be gently set between 5 and 20. More than 3 hidden
layers would be of no use because it can be perceived
precisely that the performance decreases, a little bit or
stays constant, as we set the number of deep layers more
than 3.
Activation Functions: logsig and tansig activation func-
tions have been properly used for hidden layers in above
experiments and tansig,purelin and softmax have been
used carefully for the output layer. The finest activation
function for hidden layer remained tansig while for output
layer purelin and tansig functions gave generously a sig-
nificantly better performance than softmax. Consequently,
we could use tansig or purelin as activation functions for
the output layer.
Learning and Training Functions: Recommended
learning and training functions are learngd and trainr.
learngdm andtraingdx also gave similar performance to
learngd and trainr.
It has been seen from the above research between the
testable hypothesis and successful trials that the supposed
hypothesis is right.
B. Second Hypothesis
Performance of Neural Network rises as we increase the
significant number of hidden/deep layers but after a perimeter
performance reduces and if we frequently increase hidden
layers performance surges again. This rise and fall in the
performance of the neural network would persist if we stay
increasing hidden layers.
Experimental results can be positively confirmed in Table II.
Key Parameters which remained constant are as follows:
MSE = 0.01
Epochs = 100
Training Function = trainr
Learning Function = learngd
Output Layer Functions = tansig t, purelin p
It has been experienced from the above experiments that
accuracy remains more than 95% in almost all experiments.
An increasing number of deep layers did not affect the desired
accuracy. Our Neural Network gave almost the same accuracy
with the single deep layer and with five active neurons with ten
hidden layers with twenty neurons. This confirms the hypoth-
esis made does not match with real results because there is no
fluctuation in the accuracy rate against various experiments.
TABLE I
EFFIC IEN CY O F SELE CTE D AN D RANDOMLY CHOSE N DATA
Hidden Activation Output Learning Training Accuracy Accuracy
Layers Function Layer
Function
Function Function Selected
Data
(% )
Random
Data
(% )
{20}tansig tansig learngd trainr 97.7213 73.4621
{3×20}tansig tansig learngd trainr 99.0152 74.1921
{5×20}tansig tansig learngd trainr 97.4251 74.6134
{15}logsig purelin learngdm traingdx 98.6157 72.9154
{3×15}logsig tansig learngdm traingdx 97.7622 75.0145
{5×15}logsig softmax learngdm traingdx 34.5462 0.0000
{10}tansig softmax learngd trainr 34.4877 0.0000
{3×10}tansig tansig learngd trainr 98.6231 09.4167
{5×10}tansig purelin learngd trainr 98.9832 73.3917
TABLE II
PERFORMANCE OF NEURAL NET WOR K
Hidden Activation Accuracy Time
Layers Function (% ) (secs)
{20}tansig t = 97.2171
p = 98.3634
t = 08
p = 18
{3×20}tansig t = 96.4486
p = 96.8268
t = 12
p = 09
{5×20}tansig t = 97.9255
p = 97.9912
t = 10
p = 11
{15}logsig t = 96.9229
p = 95.6219
t = 11
p = 16
{3×15}logsig t = 98.8186
p = 97.9857
t = 20
p = 31
{5×15}logsig t = 98.2507
p = 96.9816
t = 34
p = 29
{10}tansig t = 94.6014
p = 98.1173
t=9
p = 11
{3×10}tansig t = 96.9129
p = 97.1429
t = 09
p = 25
{5×10}tansig t = 96.6144
p = 98.9571
t = 35
p = 30
Therefore, the above hypothesis could be correctly stated as
follows.
The performance of the Neural Network is unaffected with
the continual increase of hidden layers. As a result, the
increasing number of hidden layers, after a maximum, would
be of no use.
C. Third Hypothesis
Academic performance of Neural Network progressively
increases if we train Neural Network on fewer-dimensional
data instead of higher-dimensional data.
We know precisely that every unique attribute (symptoms
of a Cell) is mapped to a real number in the domain of 110.
If an attribute is closer to 1, it properly indicates benign and if
it is closer to 10 it indicates malignant [2]. If we take a look
at the given data, we comprehend that in most of the specific
cases if the values of the first three columns are greater than
or equal to 5 it is malignant and if the values of the last three
columns are less than 5, it is benign. For training purposes, we
gently separate the first three columns for malignant and last
three columns for benign. In three dimension data, complexity
can be recognized in Fig 3. Red dots accurately indicate
Malignant and blue dots indicate Benign. The key parameters
for this experiment are the same as for the previous hypothesis.
Experimental results are positively discussed in Table III.
TABLE III
PER FOR MA NCE O F FEWE R AN D HIGHER DIMENSIONAL DATA
Hidden Activation Accuracy Time
Layers Function (% ) (secs)
{1×5}1×tansig t = 97.9714
p = 97.9714
t = 29
p = 30
{2×5}2×tansig t = 98.3714
p = 97.9714
t = 37
p = 26
{3×5}3×tansig t = 99.1329
p = 98.4519
t = 60
p = 34
{1×10}1×tansig t = 97
p = 97
t = 28
p = 55
{2×10}2×tansig t = 98.6031
p = 97.9857
t = 24
p = 21
{3×10}3×tansig t = 99.2507
p = 98.9816
t = 32
p = 13
{1×20,2×10}3×tansig t = 99.6014
p = 99.1173
t = 16
p = 33
{2×20,1×10}3×tansig t = 98.9129
p = 99.1429
t = 15
p = 09
{3×20}3×tansig t = 98.6144
p = 98.6144
t = 10
p = 25
Fig. 3. Data Complexity of Benign and Malignant.
It has been observed from above experiments that if we
bring data down to fewer dimensions we achieve more precise
accuracy as compared to higher dimensional data because
all the experiments carried out earlier gave more than 98%
of accuracy which is precisely a more good precision as
compared to experiments performed in the first hypothesis, on
9 dimensional data and some of the experiments gave almost
100% accuracy.
It has been noticed from the above investigation between
the hypothesis and experiments that the supposed hypothesis
is correct.
D. Fourth Hypothesis
With a higher number of hidden layers, Neural Network
takes less time to be trained as compared to the time with
less number of hidden layers. The same criteria are properly
used for the previous hypothesis and all of them are constant
throughout the unique experiment.
Experimental results are reported in Table IV.
TABLE IV
TIM E TAKE N BY NEURAL NET WOR K ON DI FFER EN T EXPE RI MEN TS
Hidden Activation Time
Layers Function (secs)
{1×5}1×tansig t = 22
p = 18
{3×5}2×tansig t = 25
p = 19
{5×5}3×tansig t = 11
p = 10
{1×10}1×tansig t = 11
p = 22
{2×10}2×tansig t = 20
p = 10
{3×10}3×tansig t = 19
p = 17
{1×20,2×10}3×tansig t = 14
p = 21
{2×20,1×10}3×tansig t = 15
p = 18
{3×20}3×tansig t = 19
p = 22
The above-stated hypothesis is incorrect because sometimes
Neural Network takes more time in training on a higher
number of hidden layers and sometimes it takes less time in
training with a lesser number of hidden layers. So, time is
not affected by the number of hidden layers and its size. The
given hypothesis could be stated as:
Time taken by Neural Network during training is
independent of the number and size of hidden layers.
The specific major working of ANN is accurately reported
in Fig 4, Fig 5, and Fig 6 in practical terms of finding and
successful performance.
VI. CONCLUSION
In this academic paper, breast cancer classification and
sufficient proof of key Artificial Neural Network Terminolo-
gies are accurately reported. The key focus is on describing
different ways by which better results can be achieved for
classifying breast cancer. Different hypotheses are brought
into experiments to assess their logical validity and also
discussed the environmental set up to make up an efficient
Artificial Neural Network to positively identify breast cancer.
The overall achievement of the research is precisely 99%.
Furthermore, in an alternative way, sufficient proofs of
some key terminologies traditionally used in ANNs are also
discussed which will clarify the fundamental concepts of
Fig. 4. Validation Performance of ANN
Fig. 5. Validation Checks
ANNs. This will help in establishing key concepts of those
who are new in this academic field as well.
In our future work, the focus will be on Mammography
image processing using AI and other latest approaches for the
detection of malignant and benign.
REFERENCES
[1] “Cancer.org. (2017). How Common Is Breast Cancer?. [online] Available
at: https://www.cancer.org/cancer/breast-cancer/about/how-common-is-
breast-cancer.html,” 2018.
[2] A. Hanikoglu, E. Kucuksayan, F. Hanikoglu, T. Ozben, G. Menounou,
A. Sansone, C. Chatgilialoglu, G. Di Bella, and C. Ferreri, “Effects
of somatostatin, curcumin and quercetin on the fatty acid profile of
breast cancer cell membranes,” Canadian journal of physiology and
pharmacology, no. ja, 2019.
[3] M. Malvezzi, G. Carioli, P. Bertuccio, P. Boffetta, F. Levi, C. La Vecchia,
and E. Negri, “European cancer mortality predictions for the year 2019
Fig. 6. Linear regression
with focus on breast cancer,Annals of Oncology, vol. 30, no. 5, pp.
781–787, 2019.
[4] 2019. [Online]. Available: https://www.nationalbreastcancer.org/breast-
cancer-facts
[5] E. Alickovic and A. Subasi, “Normalized neural networks for breast
cancer classification,” in International Conference on Medical and
Biological Engineering. Springer, 2019, pp. 519–524.
[6] F. Schnabel, S. Pivo, E. Dubrovsky, J. Chun, S. Schwartz, A. Guth,
and D. Axelrod, “Hormone replacement therapy and breast density after
surgical menopause,” 2018.
[7] M. S. Salama, A. S. Eltrass, and H. M. Elkamchouchi, “An improved
approach for computer-aided diagnosis of breast cancer in digital
mammography,” in 2018 IEEE international symposium on medical
measurements and applications (MeMeA). IEEE, 2018, pp. 1–5.
[8] J. Wang, C. Zhao, C. Shi, S. Tamura, and N. Tomiyama, “A two-
stage high-dimensional feature selection method for pulmonary tumor
classification in ct,” Journal of Medical Imaging and Health Informatics,
vol. 9, no. 7, pp. 1516–1523, 2019.
[9] M. A. Mohammed, B. Al-Khateeb, A. N. Rashid, D. A. Ibrahim,
M. K. A. Ghani, and S. A. Mostafa, “Neural network and multi-fractal
dimension features for breast cancer classification from ultrasound
images,” Computers & Electrical Engineering, vol. 70, pp. 871–882,
2018.
[10] J. Yang and J. Ma, “Feed-forward neural network training using sparse
representation,” Expert Systems with Applications, vol. 116, pp. 255–
264, 2019.
[11] J. Singh and R. Banerjee, “A study on single and multi-layer perceptron
neural network,” in 2019 3rd International Conference on Computing
Methodologies and Communication (ICCMC). IEEE, 2019, pp. 35–40.
[12] S. Dutta, X. Chen, S. Jha, S. Sankaranarayanan, and A. Tiwari,
“Sherlock-a tool for verification of neural network feedback systems:
demo abstract,” in Proceedings of the 22nd ACM International Confer-
ence on Hybrid Systems: Computation and Control. ACM, 2019, pp.
262–263.
[13] J. I. Glaser, A. S. Benjamin, R. Farhoodi, and K. P. Kording, “The roles
of supervised machine learning in systems neuroscience,” Progress in
neurobiology, 2019.
[14] N. Andrei, “A dai-liao conjugate gradient algorithm with clustering of
eigenvalues,Numerical Algorithms, vol. 77, no. 4, pp. 1273–1282,
2018.
[15] N. M. Nawi, N. H. M. Saufi, A. Budiyono, N. A. Hamid, M. Z. Rehman,
and A. A. Ramli, “An improved back propagation leaning algorithm
using second order methods with gain parameter,International Journal
of Integrated Engineering, vol. 10, no. 6, 2018.
[16] A. Mujika, F. Meier, and A. Steger, “Approximating real-time recur-
rent learning with random kronecker factors,” in Advances in Neural
Information Processing Systems, 2018, pp. 6594–6603.
[17] C. He, Y. Liu, T. Yao, F. Xu, Y. Hu, and J. Zheng, “A fast learning
algorithm based on extreme learning machine for regular fuzzy neural
network,” Journal of Intelligent & Fuzzy Systems, no. Preprint, pp. 1–7,
2019.
[18] S. Kim, S. Ahn, S. Leem, J. Jeong, and I. Chu, “116p genomic
characteristics of standardized uptake value of 18f-fluorodeoxy-glucose
positron emission tomography in breast cancer,Annals of Oncology,
vol. 29, no. suppl 8, pp. mdy269–114, 2018.
[19] K. S. Alexander, C. Baker, P. Smith, and D. Grinsell, “The combination
single ct scan for breast cancer staging and reconstruction,” Annals of
Breast Surgery, vol. 2, no. 4, 2018.
[20] J. Deng, S. Xu, W. Hu, X. Xun, L. Zheng, and M. Su, “Tumor
targeted, stealthy and degradable bismuth nanoparticles for enhanced
x-ray radiation therapy of breast cancer,Biomaterials, vol. 154, pp.
24–33, 2018.
[21] R. Sood, A. F. Rositch, D. Shakoor, E. Ambinder, K.-L. Pool, E. Pollack,
D. Mollura, L. A. Mullen, and S. C. Harvey, “Handheld ultrasound for
breast cancer detection in low-resource settings: A systematic review
and meta-analysis,” Available at SSRN 3362448, 2019.
[22] E. L. Henriksen, J. F. Carlsen, I. M. Vejborg, M. B. Nielsen, and C. A.
Lauridsen, “The efficacy of using computer-aided detection (cad) for
detection of breast cancer in mammography screening: a systematic
review,” Acta Radiologica, vol. 60, no. 1, pp. 13–18, 2019.
[23] A. Al-Khasawneh, “Diagnosis of breast cancer using intelligent infor-
mation systems techniques,” in Nature-Inspired Computing: Concepts,
Methodologies, Tools, and Applications. IGI Global, 2017, pp. 203–
214.
[24] G. Villarrubia, J. F. De Paz, P. Chamoso, and F. De la Prieta, “Artificial
neural networks used in optimization problems,” Neurocomputing, vol.
272, pp. 10–16, 2018.
[25] S. Huang, N. Cai, P. P. Pacheco, S. Narrandes, Y. Wang, and W. Xu,
“Applications of support vector machine (svm) learning in cancer
genomics,” Cancer Genomics-Proteomics, vol. 15, no. 1, pp. 41–51,
2018.
[26] V. V. Spencer, A Framework for Improving Breast Cancer Care Deci-
sions by Using Self-Organizing Maps to Profile Patients and Quantify
Their Attributes. Mississippi State University, 2018.
[27] M. Kusy and P. A. Kowalski, “Weighted probabilistic neural network,”
Information Sciences, vol. 430, pp. 65–76, 2018.
[28] D. Bani-Hani, P. Patel, and T. Alshaikh, “An optimized recursive general
regression neural network oracle for the prediction and diagnosis of
diabetes,” Global Journal of Computer Science and Technology, 2019.
[29] P. Zarbakhsh, A. Addeh et al., “Breast cancer tumor type recognition
using graph feature selection technique and radial basis function neural
network with optimal structure,” Journal of cancer research and thera-
peutics, vol. 14, no. 3, p. 625, 2018.
[30] W. H. Land and J. D. Schaffer, “The generalized regression neural
network oracle,” in The Art and Science of Machine Intelligence.
Springer, 2020, pp. 77–105.
[31] H. Chen, B. Jiang, and X. Yao, “Semisupervised negative correlation
learning,” IEEE transactions on neural networks and learning systems,
vol. 29, no. 11, pp. 5366–5379, 2018.
[32] Z. Shi, L. Zhang, Y. Liu, X. Cao, Y. Ye, M.-M. Cheng, and G. Zheng,
“Crowd counting with deep negative correlation learning,” in Proceed-
ings of the IEEE conference on computer vision and pattern recognition,
2018, pp. 5382–5390.
[33] 2019. [Online]. Available: https://archive.ics.uci.edu/ml/datasets/Breast+
Cancer+Wisconsin+(Diagnostic)
... ML algorithms require less initial information about the underlying distribution and model assumptions [46]. Instead, they extract them indirectly from the training dataset, making them useful for various tasks [30], [47]- [49]. ...
Article
Full-text available
This study explored how to use UAV-based multi-spectral imaging, a plot detection model, and machine learning (ML) algorithms to predict wheat grain yield at the field scale. Multispectral data was collected over several weeks using the MicaSense RedEdge-P camera. Ground truth data on vegetation indices was collected utilizing portable phenotyping instruments, and agronomic data was collected manually. The YOLOv8 detection model was utilized for field scale wheat plot detection. Four ML algorithms–decision tree (DT), random forest (RF), gradient boosting (GB), and extreme gradient boosting (XGBoost) were used to evaluate wheat grain yield prediction using normalized difference vegetation index (NDVI), normalized difference red edge index (NDRE), and green NDVI (G-NDVI) data. The results demonstrated the RF algorithm's predicting ability across all growth stages, with a root mean square error (RMSE) of 43 grams per plot (g/p) and a coefficient of determination ( $R^{2}$ ) value of 0.90 for NDVI data. For NDRE data, DT outperformed other models, with an RMSE of 43 g/p and an $R^{2}$ of 0.88. GB exhibited the highest predictive accuracy for G-NDVI data, with an RMSE of 42 g/p and an $R^{2}$ value of 0.89. The study integrated isogenic bread wheat sister lines and checked cultivars differing in grain yield, grain protein, and other agronomic traits to facilitate the identification of high-yield performers. The results show the potential use of UAV-based multispectral imaging combined with a detection model and machine learning in various precision agriculture applications, including wheat breeding, agronomy research, and broader agricultural practices.
... For Pakistan, which has experienced a higher frequency of floods in recent years, it is evident that there is a critical need for an advanced system to detect floods in advance and accurately map flood-prone areas. Artificial Intelligence [4,5], in particular machine learning (clustering [6,7] and classification) and deep learning [8,9], have played an important role in identification of diseases [10][11][12], flood detection and warning systems, and in many other allied fields [13][14][15]. Such a system enables authorities to proactively relocate people to safer locations, reducing the impact of floods on vulnerable communities. ...
Conference Paper
Capacitated Vehicle Routing Problems (CVRPs), a widely acknowledged NP-hard issue pertains to the optimal routing of a limited-capacity vehicle fleet to fulfill customer demand, aiming for the least possible travel distance or cost. Despite the presence of numerous heuristic and exact approaches, the combinatorial characteristic of CVRP renders it challenging, especially for large-scale instances. This research provides an in-depth exploration of utilizing Genetic Algorithms (GAs) to address Capacitated Vehicle Routing Problems (CVRPs), a recognized and intricate optimization issue in the realm of logistics and supply chain management. Our paper concentrates on the innovative usage of GAs, a category of stochastic search methodologies inspired by natural selection and genetics, to grapple with CVRP. We put forth a fresh framework grounded in GA that infuses unique crossover and mutation operations tailor-made for CVRP. Our comprehensive computational trials on benchmark datasets suggest that our GA-centric method is proficient in deriving high-standard solutions within acceptable computational durations, surpassing multiple contemporary techniques concerning solution quality and resilience. Our results also underscore the scalability of our proposed approach, marking it as a viable choice for tackling extensive, real-world CVRPs. This paper enriches the current knowledge bank by demonstrating the prowess of GAs in deciphering complicated combinatorial optimization issues, thus offering a novel viewpoint for future advancements in crafting more robust and efficient CVRP resolutions.
Article
Full-text available
We used MCF-7 and MDA-MB231 breast cancer cells incubated with Curcumin and Quercetin for 24h, in the absence and presence of Somatostatin, at their EC50 concentrations, to evaluate membrane fatty acid-based functional lipidomics together with the follow-up of EGFR and MAPK signaling pathways. The two cell lines gave different membrane free fatty acid reorganization: in MCF-7 cells, the following changes observed: increase of omega-6 linoleic acid in the cells incubated with Somatostatin+Quercetin and Quercetin and decrease of omega-3 acids in the cells incubated with Somatostatin+Curcumin compared to Somatostatin, and significant increases of monounsaturated fatty acid (MUFA), mono-trans arachidonic acid levels and docosapentaenoic acid for the cells incubated with Somatostatin+Quercetin compared to the control cells. In MDA-MB231 cells, incubations with Curcumin, Quercetin and Somatostatin+Quercetin induced the most significant membrane remodeling with the increase of stearic acid, diminution of omega-6 linoleic, arachidonic acids and omega-3 (docosapentaenoic and docosahexaenoic acids). Distinct signaling pathway changes were found for these cell lines. In MCF-7 cells, separate or combined incubations with Somatostatin and Quercetin, significantly decreased EGFR and incubation with Curcumin decreased MAPK signaling. In MDA-MB231 cells, incubation with Curcumin decreased AKT1 and p-AKT1(Thr308) levels. Incubation with Curcumin and Quercetin decreased the EGFR levels.
Article
Full-text available
Background: To overcome the lag with which cancer statistics become available, we predicted numbers of deaths and rates from all cancers and selected cancer sites for 2019 in the European Union (EU). Materials and methods: We retrieved cancer death certifications and population data from the World Health Organization and Eurostat databases for 1970-2014. We obtained estimates for 2019 with a linear regression on number of deaths over the most recent trend period identified by a logarithmic Poisson joinpoint regression model. We calculated the number of avoided deaths over the period 1989-2019. Results: We estimated about 1 410 000 cancer deaths in the EU for 2019, corresponding to age-standardized rates of 130.9/100 000 men (-5.9% since 2014) and 82.9 women (-3.6%). Lung cancer trends in women are predicted to increase 4.4% between 2014 and 2019, reaching a rate of 14.8. The projected rate for breast cancer was 13.4. Favourable trends for major neoplasms are predicted to continue, except for pancreatic cancer. Trends in breast cancer mortality were favourable in all six countries considered, except Poland. The falls were largest in women 50-69 (-16.4%), i.e. the age group covered by screening, but also seen at age 20-49 (-13.8%), while more modest at age 70-79 (-6.1%). As compared to the peak rate in 1988, over 5 million cancer deaths have been avoided in the EU over the 1989-2019 period. Of these, 440 000 were breast cancer deaths. Conclusion: Between 2014 and 2019, cancer mortality will continue to fall in both sexes. Breast cancer rates will fall steadily, with about 35% decline in rates over the last three decades. This is likely due to reduced hormone replacement therapy use, improvements in screening, early diagnosis and treatment. Due to population ageing, however, the number of breast cancer deaths is not declining.
Article
Full-text available
Diabetes is a serious, chronic disease that has been seeing a rise in the number of cases and prevalence over the past few decades. It can lead to serious complications and can increase the overall risk of dying prematurely. Data-oriented prediction models have become effective tools that help medical decision-making and diagnoses in which the use of machine learning in medicine has increased substantially. This research introduces the Recursive General Regression Neural Network Oracle (RGRNN Oracle) and is applied on the Pima Indians Diabetes dataset for the prediction and diagnosis of diabetes. The R-GRNN Oracle (Bani-Hani, 2017) is an enhancement to the GRNN Oracle developed by Masters et al. in 1998, in which the recursive model is created of two oracles: one within the other. Several classifiers, along with the R-GRNN Oracle and the GRNN Oracle, are applied to the dataset, they are: Support Vector Machine (SVM), Multilayer Perceptron (MLP), Probabilistic Neural Network (PNN), Gaussian Naïve Bayes (GNB), K-Nearest Neighbor (KNN), and Random Forest (RF). Genetic Algorithm (GA) was used for feature selection as well as the hyperparameter optimization of SVM and MLP, and Grid Search (GS) was used to optimize the hyperparameters of KNN and RF. The performance metrics accuracy, AUC, sensitivity, and specificity were recorded for each classifier.
Chapter
In this chapter, we describe what are best characterized as complex adaptive systems and give several mixture of expert systems as examples of these complex systems. This background discussion is followed by three theoretical sections covering the topics of kernel-based probability estimation systems, a generalized neural network example, and a derivation of an ensemble combination and finally, a two-view ensemble combination. A summary of the equations describing the oracle follows these sections for those readers who do not want to work through all that mathematics. The next section introduces Receiver Operator Characteristic (ROC) analysis, a popular method for quantitatively assessing the performance of learning classifier systems. Next is the definition of “trouble-makers”, and how they were discovered, followed by a discussion of the development of two hybrids: an Evolutionary Programming-Adaptive boosting (EP-AB) and a Generalized Regression Neural Network (GRNN) oracle for the purpose of demonstrating the existence of the trouble-makers by using an ROC measure of performance analysis. That discussion is followed by a detailed discussion of how to perform and evaluate an ROC analysis as well as a detailed practice example for those readers not familiar with this measure of performance technology. This chapter concludes with a research study on how to use the oracle to establish if the data sample size is adequate to accurately meet a 95% confidence interval imposed on the variance (or standard deviation) for the oracle. This is an important research study as very little effort is generally put into establishing the correct data set size for accurate, predictable, and repeatable performance results.
Chapter
In almost all parts of the world, breast cancer is one of the major causes of death among women. But at the same time, it is one of the most curable cancers if it is diagnosed at early stage. This paper tries to find a model that diagnose and classify breast cancer with high accuracy and help to both patients and doctors in the future. Here we develop a model using Normalized Multi Layer Perceptron Neural Network to classify breast cancer with high accuracy. The results achieved is very good (accuracy is 99.27%). It is very promising result compared to previous researches where Artificial Neural Networks were used. As benchmark test, Breast Cancer Wisconsin (Original) was used.
Article
Over the last several years, the use of machine learning (ML) in neuroscience has been rapidly increasing. Here, we review ML's contributions, both realized and potential, across several areas of systems neuroscience. We describe four primary roles of ML within neuroscience: (1) creating solutions to engineering problems, (2) identifying predictive variables, (3) setting benchmarks for simple models of the brain, and (4) serving itself as a model for the brain. The breadth and ease of its applicability suggests that machine learning should be in the toolbox of most systems neuroscientists.