Conference PaperPDF Available

Impact of Varying Neurons and Hidden Layers in Neural Network Architecture for a Time Frequency Application

Authors:

Abstract and Figures

In this paper, an experimental investigation is presented, to know the effect of varying the number of neurons and hidden layers in feed forward back propagation neural network architecture, for a time frequency application. Varying the number of neurons and hidden layers has been found to greatly affect the performance of neural network (NN), trained via various blurry spectrograms as input over highly concentrated time frequency distributions (TFDs) as targets, of the same signals. Number of neurons and hidden layers are varied during training and the impact is observed over test spectrograms of unknown multi component signals. Entropy and mean square error (MSE) is the decision criteria for the most optimum solution.
Content may be subject to copyright.
Impact of Varying Neurons and Hidden Layers in Neural Network
Architecture for a Time Frequency Application
Imran Shafi1, Jamil Ahmad2, MIEEE, Syed Ismail Shah2, Sr. MIEEE, and Faisal M Kashif3
1Centre for Advance Studies in Engineering, Islamabad, Pakistan
Email: imran.shafi@gmail.com
2Iqra University Islamabad Campus, H-9 Islamabad, Pakistan
Email :{ jamil,ismail}@iqraisb.edu.pk
3Laboratory for Information and Decision Systems, Massachusetts Institute of
Technology, Cambridge MA 02139, USA
Email: fmkashif@mit.edu
Abstract
In this paper, an experimental investigation is presented,
to know the effect of varying the number of neurons and
hidden layers in feed forward back propagation Neural
Network Architecture, for a Time Frequency application.
Varying the number of neurons and hidden layers has
been found to greatly affect the performance of Neural
Network (NN), trained via various blurry spectrograms as
input over highly concentrated Time Frequency
Distributions (TFDs) as targets, of the same signals.
Number of neurons and hidden layers are varied during
training and the impact is observed over test spectrograms
of unknown multi component signals. Entropy and Mean
Square Error (MSE) is the decision criteria for the most
optimum solution.
Key words: Neural Networks, Back propagation, hidden
layer, Time Frequency Analysis, Neurons
1. Introduction
The brain is a very efficient tool. Having about 100,000
times slower response time than computer chips, it (so far)
beats the computer in complex tasks, such as image and
sound recognition, motion control and so on. It is also
about 10,000,000,000 times more efficient than the
computer chip in terms of energy consumption per
operation. An Artificial Neural Network (ANN) is an
information processing paradigm that is inspired by the
way, the brain process information [1]. The key element of
this paradigm is the novel structure of the information
processing system. It is composed of a large number of
highly interconnected processing elements (neurons)
working in unison to solve specific problems. ANNs, like
people, learn by example. An ANN is configured for a
specific application, such as pattern recognition or data
classification, through a learning process. Learning in
biological systems involves adjustments to the synaptic
connections that exist between the neurons. This is true of
ANNs as well [6].
1.1 Human Vs Artificial Neuron. A typical human
neuron collects signals from others through a host of fine
structures called dendrites. The neuron sends out spikes of
electrical activity through a long, thin stand known as an
axon, which splits into thousands of branches. At the end
of each branch, a structure called a synapse converts the
activity from the axon into electrical effects that inhibit or
excite activity from the axon into electrical effects that
inhibit or excite activity in the connected neurons. When a
neuron receives excitatory input that is sufficiently large
compared with its inhibitory input, it sends a spike of
electrical activity down its axon. Learning occurs by
changing the effectiveness of the synapses so that the
influence of one neuron on another changes.
The essential features of human’s neurons and their
interconnections are estimated. We then typically program
a computer to simulate these features. However because
our knowledge of neurons is incomplete and our
computing power is limited, our models are necessarily
gross idealizations of real networks of neurons. A model
of human’s neuron Vs Artificial Neuron is presented in
figure 3.
1.2 ANN Layers. The commonest type of ANN
consists of three groups, or layers, of units: a layer of
"input" units is connected to a layer of "hidden" units,
which is connected to a layer of "output" units. The
activity of the input units represents the raw information
that is fed into the network. The activity of each hidden
unit is determined by the activities of the input units and
the weights on the connections between the input and the
hidden units. The behavior of the output units depends on
the activity of the hidden units and the weights between
the hidden and output units.
This simple type of network is interesting because the
hidden units are free to construct their own representations
of the input. The weights between the input and hidden
units determine when each hidden unit is active, and so by
modifying these weights, a hidden unit can choose what it
represents. We also distinguish single-layer and multi-
layer architectures. The single-layer organization, in which
all units are connected to one another, constitutes the most
general case and is of more potential computational power
than hierarchically structured multi-layer organizations. In
multi-layer networks, units are often numbered by layer,
instead of following a global numbering.
The most widely used architecture in ANNs has been
the Multiple Layer Perceptron (MLP), trained with the
Back Propagation (BP) error learning algorithm. However,
the MLP suffers from fundamental problems like
convergence time, local minima and absence of a simple
rule to obtain the right number of neurons and hidden
layers.
In this paper we have used Feed forward Back
Propagation NN to find the solution of the last problem.
An ANN is trained with various blurry spectrograms as
input to be mapped over highly concentrated TFDs of
same signals [2]. Test spectrograms of multi component
signals are then presented to trained NN. Optimum
solution is explored for the application as far as number of
neurons and hidden layers are concerned, to get the best
concentration along Instantaneous Frequencies (IFs) for
resultant images. The concentration is measured in terms
of entropies [4] of the resultant TFDs. The lower the
entropy, higher is the concentration along IF. In this paper
entropy [4] of (, )Qn
ω
is considered as measure of
concentration given by:
() ()
1
2
0,log , 0
N
Qn
EQnQnd
π
π
ωωω
=
=− ≥
;
(1)
Rest of the paper is organized as follows. Section 2 & 3
describes the NN architecture and procedural detail.
Section 4 covers the simulation results and Section 5
concludes the paper.
2. The NN Architecture
The brain basically learns from experience. NNs are
sometimes called machine learning algorithms, because
changing of its connection weights (training) causes the
network to learn the solution to a problem. The strength of
connection between the neurons is stored as a weight-
value for the specific connection. The system learns new
knowledge by adjusting these connection weights. The
learning ability of a neural network is determined by its
architecture and by the algorithmic method chosen for
training.
2.1 How BP Algorithm works?
In order to train a NN to perform some task, we must
adjust the weights of each unit in such a way that the error
between the desired output and the actual output is
reduced. This process requires that the neural network
compute the error derivative of the weights (EW). In other
words, it must calculate how the error changes as each
weight is increased or decreased slightly. The BP
algorithm is the most widely used method for determining
the EW.
The BP algorithm is easiest to understand if all the
units in the network are linear. The algorithm computes
each EW by first computing the EA, the rate at which the
error changes as the activity level of a unit is changed. For
output units, the EA is simply the difference between the
actual and the desired output. To compute the EA for a
hidden unit in the layer just before the output layer, we
first identify all the weights between that hidden unit and
the output units to which it is connected. We then multiply
those weights by the EAs of those output units and add the
products. This sum equals the EA for the chosen hidden
unit. After calculating all the EAs in the hidden layer just
before the output layer, we can compute in like fashion the
EAs for other layers, moving from layer to layer in a
direction opposite to the way activities propagate through
the network. This is what gives back propagation its name.
Once the EA has been computed for a unit, it is straight
forward to compute the EW for each incoming connection
of the unit. The EW is the product of the EA and the
activity through the incoming connection.
2.1.1 Various Steps. The BP algorithm consists of four
steps:
1. Compute how fast the error changes as the activity of an
output unit is changed. This error derivative (EA) is the
difference between the actual and the desired activity.
(2)
2. Compute how fast the error changes as the total input
received by an output unit is changed. This quantity (EI) is
the answer from step 1 multiplied by the rate at which the
output of a unit changes as its total input is changed.
(3)
3. Compute how fast the error changes as a weight on the
connection into an output unit is changed. This quantity
(EW) is the answer from step 2 multiplied by the activity
level of the unit from which the connection emanates.
(4)
4. Compute how fast the error changes as the activity of a
unit in the previous layer is changed. This crucial step
allows back propagation to be applied to multilayer
networks. When the activity of a unit in the previous layer
changes, it affects the activites of all the output units to
which it is connected. So to compute the overall effect on
the error, we add together all these seperate effects on
output units. But each effect is simple to calculate. It is the
answer in step 2 multiplied by the weight on the
connection to that output unit.
(5)
By using steps 2 and 4, we can convert the EAs of one
layer of units into EAs for the previous layer. This
procedure can be repeated to get the EAs for as many
previous layers as desired. Once we know the EA of a
unit, we can use steps 2 and 3 to compute the EWs on its
incoming connections.
2.2 NN Topology
In this paper grayscale blurry TFDs are considered and
Levenberg-Marquardt Back propagation (LMB) training
algorithm with feed forward back propagation architecture
is used. No of hidden layer and neurons are varied to find
the optimum solution. The ‘tansig’ and ‘poslin’ transfer
functions are used in between input-hidden layers and
hidden-output layers respectively. Multiple layers of
neurons with nonlinear transfer functions allow the
network to learn nonlinear and linear relationships
between input and output vectors. The linear output layer
lets the network produce values outside the range -1 to +1.
3. The Procedural Details
Here we have targeted a Time Frequency application to
find the most optimum topology/architecture of NN. To
achieve the objective, we proceeded as under:
a. Number of sub spaces are decided for clustering the
available data.
b. We select normalized sub space direction vectors that
will best represent the subspaces. The directional
vectors are used to characterize different types of
edges in the image. The choice is dictated by the
problem of deblurring. Here are few issues that are
considered:
(1) Edges are important image characteristics.
(2) Blurring results in loss of edge information
from images.
(3) The process of deblurring may produce a more
useful image.
c. Input data is vectorized and correlation between each
input vector and directional vectors is calculated to
assign it to the correct subspace. This creates a certain
clustering effect on the input vectors since a vector
will lie in the subspace represented by directional
vector that is most similar to this vector with respect
to its information content.
d. For each cluster, NNs are trained by varying the
nunber of neurons and hidden layers. Test TFDs are
given to the trained NNs to find the most optimum
solution in terms of number of neurons and hidden
layers for the application under consideration. Best
topology/architecture is finalized on the basis of
performance measured as entropies of the resultant
images.
3.1 Training/Test TFDs
To train the NNs with algorithm described earlier, the
spectrogram of the two parallel chirps signal is used as
input. The grayscale spectrogram of this signal is shown as
figure 4. The respective target time-frequency plane image
of same signal is shown in figure 4.
510 15 20 25 30 35 40 45 50
0
0.01
0.02
0.03
0.04
0.05
0.06
no of neuron
error converged
ERROR VS NO OF NEURONS
0 5 10 15 20 25 30 35 40
10
-4
10
-3
10
-2
10
-1
10
0
10
1
epoches
MSE
ERROR VS EP OCHES
3 LAYERS
2 LAYERS
1 LAYERS
3.1.1 Parallel Chirps Signal It is given by:
(
)
(
)
(
)
12
nxnxnY=+ (5) (5)
Where Where
()
()
1
1with
()
1/4nn
π
ω
= and
()
()
2
2
j
nn
nex
ω
=with
()
234
n
nN
π
π
ω
=+ ;
()
()
1
1
nn
nex
ω
=N
Here N represents the total number of points in the signal.
3.1.2 Test Signal we have fed spectrogram of single
chirp signal as test image (figure 5) to the trained NN.
Discussion of experimental results is presented in next
section.
4. Simulation Results
There are a lot of factors that affect the performance of
NN such as number of hidden layers, neurons in the hidden
layer, learning rate and momentum term etc. We have carried
out simulation for the first two factors which are presented
below:
4.1 Effect of number of neurons in the hidden layers
We have studied the effect of the number of neurons in
the hidden layer. The network was tested with
2,3,4,5,10,15,20,30,40 and 50 neurons in single/multiple
hidden layer(s). The network never converged to a stable
point when we tried the network with neurons upto 30. The
reason being that the less number of neurons take the data
from the input grid, and hence fails to convey the correct
information to the next layers. The results were satisfactory
with 35 neurons in the hidden layer, but by increasing the
number of neurons further, no improvement was observed in
the reduction of error in last epoch as shown in figure 1.
Entropy values are also minimum for 40 or more neurons
irrespective of number of hidden layers, as given in table I.
4.2 Effect of number of hidden layers
Number of hidden layers is the most important criteria
while studying the architecture of the NNs. We have varied
the number of hidden layers for the given input sets and the
results are shown in figures 6 to 12. It is noted that the result
even deteriorated if we use more then single hidden layer.
Our study also verifies that the complex non linear problem
at hand can be solved with single hidden layer, so there is no
significant need for 2 or more layer architecture. The same
fact is strengthened by the entropy values mentioned in table
I.
Figure 1: Error Vs Number of neurons in single hidden
layers
Figure 2: Error Vs epochs performance for various number
of hidden layers
5. Conclusions
The simulation results presented in the paper indicated
that NN architecture composed of single hidden layer with
40 neurons is able to remove the blur from the unknown
spectrograms effectively with minimum MSE in last epoch
and lowest entropy values as given in Table I. Increasing
the number of neurons/hidden layers further only seems to
increase the complexity of the network, and is found to be
unsuitable manifested by both visual (figures 6-12) and
mathematical findings (Table I). Studying the effect of
these parameters in other applications will be a major
work for future research.
6. References
[1] K. Jain, J. Mao and K. M. Mohiddin, “Artificial Neural
Network: A tutorial”, IEEE Trans. on Computers, pp. 31-44,
1996.
[2] I. Shafi, J. Ahmad, S.I. Shah, FM. Kashif, “ Evolutionary De-
noised and Concentrated Time Frequency Distributions (TFDs)
using Bayesian Regularized Neural Network Model”, Under
Review at Journal of IEEE Transactions on Neural Networks, 2nd
Draft submitted on 21 Aug 2006.
[3] I. Shafi, J. Ahmad, S.I. Shah, FM. Kashif, “ Time Frequency
Distribution using Neural Networks”, Proceeding of IEEE
International Conf on Emerging Technologies, pp. 32-35,
Pakistan, 2005.
[4] R.M. Gray, “Entropy and Information Theory”. New York
Springer-Verlag, 1990.
[5] L. Cohen, “Time Frequency Analysis”, Prentice-Hall, NJ,
1995.
[6] M.T. Hagan, H.B. Demuth & M. Beale, “Neural Network
Design”, Thomson Learning USA, 1996.
[7] J. Ahmad, I. Shafi, S.I. Shah, FM. Kashif, “Analysis and
Comparison of Neural Network Training Algorithms for the Joint
Time-Frequency Analysis”, Proceeding of IASTED International
Conf on Artificial Intelligence and application, pp. 193-198,
Austria, Feb 2006.
Figure 3: (a) Human Neuron (b) Artificial Neuron
Figure 4: Input training/target images of parallel chirps
signal
Figure 5: Test image of single chirp signal
Figure 6: Resultant image with 2 layers, 50 neurons
(
a
)
(
b
)
Figure 7: Resultant image with 2 layers 5 neurons
Figure 8: Resultant image with 3 layers 5 neurons
Figure 9: Resultant image with 3 layers 20 neurons
Figure 10: Resultant image with 2 layer 15 neurons
Figure 11: Resultant image with 1 layer 20 neurons
Figure 12: Resultant image with single hidden layer
having 40 neurons
TABLE I
Impact of varying neurons and hidden layers over
entropy of resultant image
Description Number of Neurons
10 20 30 40 50
Q
E
bits for single
layer
10.20 9.31 9.01 8.20 8.20
Q
E
bits for 2
layers
20.41 16.31 15.60 11.21 11.21
Q
E
bits for 3
layer
22.56 14.30 12.10 10.24 10.24
... Additionally, adding more hidden layers can increase the network's computational complexity and training time, which can be a disadvantage in some applications where speed is important. The best number of hidden layers depends on the task specification and dataset and may require some experimentation and tuning [94,95]. ...
Article
In the last few years, Neural Networks have become more common in different areas due to their ability to learn intricate patterns and provide precise predictions. Nonetheless, creating an efficient neural network model is a difficult task that demands careful thought of multiple factors, such as architecture, optimization method, and regularization technique. This paper aims to comprehensively overview the state-of-the-art artificial neural network (ANN) generation and highlight key challenges and opportunities in machine learning applications. It provides a critical analysis of current neural network model design methodologies, focusing on the strengths and weaknesses of different approaches. Also, it explores the use of different deep neural networks (DNN) in image recognition, natural language processing, and time series analysis. In addition, the text explores the advantages of selecting optimal values for various components of an Artificial Neural Network (ANN). These components include the number of input/output layers, the number of hidden layers, the type of activation function used, the number of epochs, and the model type selection. Setting these components to their ideal values can help enhance the model's overall performance and generalization. Furthermore, it identifies some common pitfalls and limitations of existing design methodologies, such as overfitting, lack of interpretability, and computational complexity. Finally, it proposes some directions for future research, such as developing more efficient and interpretable neural network architectures, improving the scalability of training algorithms, and exploring the potential of new paradigms, such as Spiking Neural Networks, quantum neural networks, and neuromorphic computing.
... Therefore, it is crucial to determine the proper number of neurons for the initial hidden layer, as it negates the need for additional hidden layers. Similarly, increasing the number of neurons or hidden layers results in making the network more complex (Shafi et al., 2006). In this study, we used the maximum available number of neurons (100) which is consistent with Uzair and Jamil (2020) study to avoid inaccuracies in complex situations due to improperly trained neural networks. ...
... According to Fig. 8, the performance measures increased as the number of neurons in each layer increased. This supports earlier studies, which reported that an increase in the number of neurons improved performance metrics (Oyebisi and Alomayri 2023a;Shafi et al. 2006). However, the metric achieved its best performance at 25 neurons compared with 30 neurons for training, validation, and testing with 99.91, 99.62, and 99.01% ...
Article
Full-text available
Precise forecast of river discharge is crucial for a variety of sectors, from human activities to the control of environmental hazards, considering growing need for water resources and the effects of climate change. Despite the development of various discharge forecasting models, real-time projections are still difficult. This has necessitated the application of Artificial Intelligence to predict river discharge using satellite data since there is paucity of gauged records in most developing countries. In this research, a 38-year data, obtained from the National Aeronautics and Space Administration (NASA)/Goddard Space Flight Center using the Modern-Era Retrospective Analysis for Research and Applications, version 2 (MERRA-2), was used to model the discharge of five selected rivers from South Eastern Nigeria watershed. Deep Neural Networks (DNN) modeling technique was engaged. Back propagation learning algorithms of various network topologies were developed for predicting the river’s discharge with respect to other hydrological properties. The developed model was trained and validated with the raw dataset. Results indicated that relative humidity, atmospheric pressure, wind speed, rainfall intensity, radiation, air temperature, and soil temperature govern the discharge of river. The DNN model accurately predicted the river discharge with the 7–25-25–25-1 network structure, as evidenced by 99.91, 99.62, and 99.01% R for the training, validation, and test. The results of this analysis showed that DNN approach is effective at forecasting river discharge with respect to the hydrological characteristics. Decision-makers in the water and environmental sectors can utilize this knowledge in making an informed sustainable development plan.
... The pursuit learning network can be acknowledged as the most famous modular structure. Results depicted that algorithm performed better for solving the regression problems [3]. Moderation concept of neurons were adopted to update the neurons on input and output of neural network architecture [4]. ...
Article
Full-text available
Identification of optimal number of neurons and layers in a proposed neural architecture is very complex for the better results. The determination of the hidden layer number is also very difficult task for the proposed network. The recognition of the effective neural network model in terms of accuracy and precision in results as well as in terms of computational resources is very crucial in the community of the computer scientists. An effective proposed neural network architecture must comprise the appropriate numbers of perceptrons and number of layers. Another research gap was also reported by the researchers community that the perceptron stuck during the training phase in finding minima or maxima for stochastic gradient to solve any engineering application. Therefore to resolve the problem of selection of neurons and layers an analysis was performed to evaluate the performance of the neural network architecture with different neurons and layers on the same data set. The results revealed that the justified network architecture would contain justified number of neurons and layers as more number of neurons and layers increase more computational resources and training time. It was suggested that a neural network architecture should be proposed comprising of minimum 2 to 5 layers. Entropy and Mean square error was considered as a yardstick to measure the neural network architecture performance. Results depicted that the an effective neural network architecture must initially be simulated or checked with minimum number of instances to evaluate the model.
Article
The study presents a novel technique for lung auscultation based on graph theory, emphasizing the potential of graph parameters in distinguishing lung sounds and supporting earlier detection of various respiratory pathologies. The frequency spread and the component magnitudes are revealed from the analysis of eighty-five bronchial (BS) and pleural rub (PS) lung sounds employing the power spectral density (PSD) plot and wavelet scalogram. The low-frequency spread, and persistence of the high-intensity frequency components are visible in BS sounds emanating from the uniform cross-sectional area of the trachea. The frictional rub between the pleurae causes a higher frequency spread of low-intensity intermittent frequency components in PS signals. From the complex networks of BS and PS, the extracted graph features are - graph density (\( G)\), transitivity (\( T)\), degree centrality (\( {D}_{c}\)), betweenness centrality (\( {C}_{b})\), eigenvector centrality (\( {E}_{c}\)), and graph entropy (En). The high values of \( G\) and \( T\) show a strong correlation between distinct segments of the BS signal originating from a consistent cross-sectional tracheal diameter and, hence, the generation of high-intense low-spread frequency components. An intermittent low-intense and a relatively greater frequency spread in PS signal appear as high \( {D}_{c}\), \( {C}_{b}\), \( {E}_{c}\), and \( {E}_{n}\) values. With these complex network parameters as input attributes, the supervised machine learning techniques– discriminant analyses, support vector machines, k-nearest neighbors, and neural network pattern recognition (PRNN)– classify the signals with more than 90% accuracy, with PRNN having 25 neurons in the hidden layer achieving the highest (98.82%).
Article
The study presents an intelligent data processing algorithm based on artificial neural networks (ANN) and adaptive neuro-fuzzy inference systems (ANFIS) to predict the dynamic behavior of Dębica railway steel arch bridge produced from dynamic responses of steel hangers during the passage of trains. Field data sets were collected from the vibration-based structural health monitoring (VSHM) system of the hangers and bridge deck over a nine-month period from December 2019 to September 2020. The input variables of the ANN and ANFIS models consist of RMS (Root Mean Square) values of vibration signals installed on hangers, and the output is RMS values of dynamic responses on each of two bridge spans. Optimizing ANN architecture based on the genetic algorithm (GA) is implemented to determine the number of neurons in the hidden layers of the ANN regression models. The results indicate that the accuracy of optimized ANN models is up to 85% of training and 79% of testing. Additionally, optimized ANN prediction models have been shown to outperform ANFIS regression models among the six proposed strategies. The results of this study offer a systematic structural diagnostic for decision-making to emerging technology in the VSHM system.
Article
Full-text available
Back propagation neural network (BPNN) is one of the most basic and commonly used models in machine learning. Hidden layers play a crucial function in maximizing the performance of neural networks, especially when solving complicated issues that demand strict adherence to accuracy and time complexity requirements. The only reliable ways at the moment are just experience and attempting each situation, as the process of determining the amount of Hidden Layer neurons is still unclear. To investigate this relationship, this article conducted extensive experiments involving designing and training the BPNN model with varying numbers of hidden layer neurons. Leverage benchmark data sets and quantify accuracy with appropriate error metrics. The analysis in this article focuses on understanding the impact of different neuron counts on network performance. Under specific assumptions, the findings show a relationship between the quantity of hidden layer neurons and BPNN accuracy. According to statistics of some recent neural network applications, it can be observed that if the number of neurons in the hidden layer is decreased, it will have an effect on the network's accuracy because complex problems with a small number of hidden layers may cause the network to be incorrectly trained; However, relative to the accuracy gain, the time complexity rises orders of magnitude as the number of hidden layer neurons exceeds the ideal amount.
Article
Full-text available
Impact copolymers are synthesized using ethylene and propylene monomers in various proportions. The mechanical properties of these polymers are directly influenced by the ethylene content because an increase in ethylene results in significant enhancements in the copolymer properties. To precisely quantify the ethylene content, ATR-FTIR was employed, leveraging calibration methodologies and prediction models based on machine learning (ML). Principal component regression (PCR), artificial neural networks (ANN), support vector regression (SVR), and k-nearest neighbor (kNN) were used. Green ethylene in concentrations ranging from 0.5% to 53% was used. R 2 , root mean square error of calibration (NRMSE), and root mean square error of prediction (RMSEP) were used as parameters for judging the best ML model. This study demonstrated that focusing on bands between 690 and 1325 cm-1 enables effective classification and prediction of green ethylene concentrations. Even when all measured bands within this range are reduced to a space of only two principal components, a remarkable 97% of the variance is explained. The results suggest that mid-infrared spectroscopy could be a useful tool for quantitative analysis of green ethylene when machine learning algorithms are used.
Conference Paper
Full-text available
In this paper we present a comparison of Neural Network Training Algorithms for obtaining a Time Frequency Distribution (TFD) of a signal whose frequency components vary with time. The method employs various algorithms used in NNs which are trained by using the spectrograms of several training signals as input and TFDs that are highly concentrated along the instantaneous frequencies (IFs) of the individual components present in the signal as targets. The trained neural networks are then presented with the spectrogram of unknown signals. We compute the entropy as a measure of the result obtained and carry out error and time analysis to compare the performance of algorithms used.
Conference Paper
Full-text available
In this paper we present a method of obtaining a time frequency distribution (TFD) of a signal whose frequency components vary with time. The method employs neural networks (NN) which are trained by using the spectrograms of several training signals as input and TFDs that are highly concentrated along the instantaneous frequencies of the individual components present in the signal as targets. The trained neural network is then presented with the spectrogram of unknown signals and highly concentrated TFDs are obtained.
Article
Full-text available
Time-frequency distributions (TFDs) that are highly concentrated in the time-frequency plane are computed using a Bayesian regularised neural network model. The degree of regularisation is automatically controlled in the Bayesian inference framework and produces networks with better generalised performance and lower susceptibility to over-fitting. Spectrograms and Wigner transforms of various known signals form the training set. Simulation results show that regularisation, with input training under Mackay's evidence framework, produces results that are highly concentrated along the instantaneous frequencies of the individual components present in the test TFDs. Various parameters are compared to establish the effectiveness of the approach.
Article
We obtain explicit expressions for the time, scale, and frequency moments of the wavelet transform in terms of the moments of the signal and wavelet. We show that generally they do not exist for common signals even when the moments of the signal and wavelet do exit. The peculiar behavior of the wavelet transform in this regard is pinpointed. The lack of existence of simple moments makes the interpretation and usefulness of the wavelet transform for time- frequency analysis problematic and it is argued that its behavior is quite poor when compared to other simple time-frequency methods, such as the short-time Fourier transform.
Book
This book is an updated version of the information theory classic, first published in 1990. About one-third of the book is devoted to Shannon source and channel coding theorems; the remainder addresses sources, channels, and codes and on information and distortion measures and their properties. New in this edition: •Expanded treatment of stationary or sliding-block codes and their relations to traditional block codes •Expanded discussion of results from ergodic theory relevant to information theory •Expanded treatment of B-processes - processes formed by stationary coding memoryless sources •New material on trading off information and distortion, including the Marton inequality •New material on the properties of optimal and asymptotically optimal source codes •New material on the relationships of source coding and rate-constrained simulation or modeling of random processes Significant material not covered in other information theory texts includes stationary/sliding-block codes, a geometric view of information theory provided by process distance measures, and general Shannon coding theorems for asymptotic mean stationary sources, which may be neither ergodic nor stationary, and d-bar continuous channels. © Springer Science+Business Media, LLC 2011. All rights reserved.
Article
Numerous advances have been made in developing intelligent programs, some inspired by biological neural networks. Researchers from many scientific disciplines are designing artificial neural networks (ANNs) to solve a variety of problems in pattern recognition, prediction, optimization, associative memory; and control. Although successful conventional applications can be found in certain well-constrained environments, none is flexible enough to perform well outside its domain. ANNs provide exciting alternatives, and many applications could benefit from using them. This article is for those readers with little or no knowledge of ANNs to help them understand the other articles in this issue of Computer. It discusses the motivation behind the development of ANNs; describes the basic biological neuron and the artificial computation model; outlines network architectures and learning processes; and presents multilayer feed-forward networks, Kohonen's self-organizing maps, Carpenter and Grossberg's Adaptive Resonance Theory models, and the Hopfield network. It concludes with character recognition, a successful ANN application.