There is an evidence that channel estimation in communication systems plays a crucial issue in recovering the transmitted data. In recent years, there has been an increasing interest to solve problems due to channel estimation and equalization especially when the channel impulse response is fast time varying Rician fading distribution that means channel impulse response change rapidly. Therefore, there must be an optimal channel estimation and equalization to recover transmitted data. However. this paper attempt to compare epsilon normalized least mean square (ε-NLMS) and recursive least squares (RLS) algorithms by computing their performance ability to track multiple fast time varying Rician fading channel with different values of Doppler frequency, as well as mean square deviation (MSD) has simulated to measure the difference between original channel and what is estimated. The simulation results of this study showed that (ε-NLMS) tend to perform fast time varying Rician fading channel better than (RLS) adaptive filter.
The - mixing ratios of -transitions from levels in populated in the reactions are calculated in present work using - ratio, constant statisticalTensor and least squares fitting methods The results obtained are in general, in good agreement or consistent, within the associated uncertainties, with these reported in Ref.[9],the discrepancies that occurs are due to inaccuracy existing in the experimental data The results obtained in the present work confirm the –method for mixed transitions better than that for pure transition because this method depends only on the experimental data where the second method depends on the pure or those considered to be pure -transitions, the same results occur in – method
Abstract
The logistic regression model is one of the nonlinear models that aims at obtaining highly efficient capabilities, It also the researcher an idea of the effect of the explanatory variable on the binary response variable. &nb
... Show MoreThe logistic regression model is an important statistical model showing the relationship between the binary variable and the explanatory variables. The large number of explanations that are usually used to illustrate the response led to the emergence of the problem of linear multiplicity between the explanatory variables that make estimating the parameters of the model not accurate.
... Show MoreTraditionally, path selection within routing is formulated as a shortest path optimization problem. The objective function for optimization could be any one variety of parameters such as number of hops, delay, cost...etc. The problem of least cost delay constraint routing is studied in this paper since delay constraint is very common requirement of many multimedia applications and cost minimization captures the need to
distribute the network. So an iterative algorithm is proposed in this paper to solve this problem. It is appeared from the results of applying this algorithm that it gave the optimal path (optimal solution) from among multiple feasible paths (feasible solutions).
Abstract
For sparse system identification,recent suggested algorithms are -norm Least Mean Square (
-LMS), Zero-Attracting LMS (ZA-LMS), Reweighted Zero-Attracting LMS (RZA-LMS), and p-norm LMS (p-LMS) algorithms, that have modified the cost function of the conventional LMS algorithm by adding a constraint of coefficients sparsity. And so, the proposed algorithms are named
-ZA-LMS,
The regressor-based adaptive control is useful for controlling robotic systems with uncertain parameters but with known structure of robot dynamics. Unmodeled dynamics could lead to instability problems unless modification of control law is used. In addition, exact calculation of regressor for robots with more than 6 degrees of freedom is hard to be calculated, and the task could be more complex for robots. Whereas the adaptive approximation control is a powerful tool for controlling robotic systems with unmodeled dynamics. The local (partitioned) approximation-based adaptive control includes representation of the uncertain matrices and vectors in the robot model as finite combinations of basis functions. Update laws for the weighting matri
... Show MoreEncryption of data is translating data to another shape or symbol which enables people only with an access to the secret key or a password that can read it. The data which are encrypted are generally referred to as cipher text, while data which are unencrypted are known plain text. Entropy can be used as a measure which gives the number of bits that are needed for coding the data of an image. As the values of pixel within an image are dispensed through further gray-levels, the entropy increases. The aim of this research is to compare between CAST-128 with proposed adaptive key and RSA encryption methods for video frames to determine the more accurate method with highest entropy. The first method is achieved by applying the "CAST-128" and
... Show More