In this paper we give definitions, properties and examples of the notion of type Ntopological space. Throughout this paper N is a finite positive number, N 2. The task of this paper is to study and investigate some properties of such spaces with the existence of a relation between this space and artificial Neural Networks (ïNN'S), that is we applied the definition of this space in computer field and specially in parallel processing
The primary objective of this research be to develop a novel thought of fibrewise micro—topological spaces over B. We present the notions from fibrewise micro closed, fibrewise micro open, fibrewise locally micro sliceable, and fibrewise locally micro-section able micro topological spaces over B. Moreover, we define these concepts and back them up with proof and some micro topological characteristics connected to these ideas, including studies and fibrewise locally micro sliceable and fibrewise locally micro-section able micro topological spaces, making it ideal for applications where high-performance processing is needed. This paper will explore the features and benefits of fibrewise locally micro-sliceable and fibrewise locally
... Show MoreThis paper presents a proposed neural network algorithm to solve the shortest path problem (SPP) for communication routing. The solution extends the traditional recurrent Hopfield architecture introducing the optimal routing for any request by choosing single and multi link path node-to-node traffic to minimize the loss. This suggested neural network algorithm implemented by using 20-nodes network example. The result shows that a clear convergence can be achieved by 95% valid convergence (about 361 optimal routes from 380-pairs). Additionally computation performance is also mentioned at the expense of slightly worse results.
This paper adapted the neural network for the estimating of the direction of arrival (DOA). It uses an unsupervised adaptive neural network with GHA algorithm to extract the principal components that in turn, are used by Capon method to estimate the DOA, where by the PCA neural network we take signal subspace only and use it in Capon (i.e. we will ignore the noise subspace, and take the signal subspace only).
Discriminant between groups is one of the common procedures because of its ability to analyze many practical phenomena, and there are several methods can be used for this purpose, such as linear and quadratic discriminant functions. recently, neural networks is used as a tool to distinguish between groups.
In this paper the simulation is used to compare neural networks and classical method for classify observations to group that is belong to, in case of some variables that don’t follow the normal distribution. we use the proportion of number of misclassification observations to the all observations as a criterion of comparison.
In Automatic Speech Recognition (ASR) the non-linear data projection provided by a one hidden layer Multilayer Perceptron (MLP), trained to recognize phonemes, and has previous experiments to provide feature enhancement substantially increased ASR performance, especially in noise. Previous attempts to apply an analogous approach to speaker identification have not succeeded in improving performance, except by combining MLP processed features with other features. We present test results for the TIMIT database which show that the advantage of MLP preprocessing for open set speaker identification increases with the number of speakers used to train the MLP and that improved identification is obtained as this number increases beyond sixty.
... Show MoreIn this paper we describe several different training algorithms for feed forward neural networks(FFNN). In all of these algorithms we use the gradient of the performance function, energy function, to determine how to adjust the weights such that the performance function is minimized, where the back propagation algorithm has been used to increase the speed of training. The above algorithms have a variety of different computation and thus different type of form of search direction and storage requirements, however non of the above algorithms has a global properties which suited to all problems.
The main aim of image compression is to reduce the its size to be able for transforming and storage, therefore many methods appeared to compress the image, one of these methods is "Multilayer Perceptron ". Multilayer Perceptron (MLP) method which is artificial neural network based on the Back-Propagation algorithm for compressing the image. In case this algorithm depends upon the number of neurons in the hidden layer only the above mentioned will not be quite enough to reach the desired results, then we have to take into consideration the standards which the compression process depend on to get the best results. We have trained a group of TIFF images with the size of (256*256) in our research, compressed them by using MLP for each
... Show MoreIn this research Artificial Neural Network (ANN) technique was applied to study the filtration process in water treatment. Eight models have been developed and tested using data from a pilot filtration plant, working under different process design criteria; influent turbidity, bed depth, grain size, filtration rate and running time (length of the filtration run), recording effluent turbidity and head losses. The ANN models were constructed for the prediction of different performance criteria in the filtration process: effluent turbidity, head losses and running time. The results indicate that it is quite possible to use artificial neural networks in predicting effluent turbidity, head losses and running time in the filtration process, wi
... Show MoreThe aim of this paper is to design fast neural networks to approximate periodic functions, that is, design a fully connected networks contains links between all nodes in adjacent layers which can speed up the approximation times, reduce approximation failures, and increase possibility of obtaining the globally optimal approximation. We training suggested network by Levenberg-Marquardt training algorithm then speeding suggested networks by choosing most activation function (transfer function) which having a very fast convergence rate for reasonable size networks. In all algorithms, the gradient of the performance function (energy function) is used to determine how to
... Show More