Krawtchouk polynomials (KPs) and their moments are promising techniques for applications of information theory, coding theory, and signal processing. This is due to the special capabilities of KPs in feature extraction and classification processes. The main challenge in existing KPs recurrence algorithms is that of numerical errors, which occur during the computation of the coefficients in large polynomial sizes, particularly when the KP parameter (p) values deviate away from 0.5 to 0 and 1. To this end, this paper proposes a new recurrence relation in order to compute the coefficients of KPs in high orders. In particular, this paper discusses the development of a new algorithm and presents a new mathematical model for computing the initial value of the KP parameter. In addition, a new diagonal recurrence relation is introduced and used in the proposed algorithm. The diagonal recurrence algorithm was derived from the existing n direction and x direction recurrence algorithms. The diagonal and existing recurrence algorithms were subsequently exploited to compute the KP coefficients. First, the KP coefficients were computed for one partition after dividing the KP plane into four. To compute the KP coefficients in the other partitions, the symmetry relations were exploited. The performance evaluation of the proposed recurrence algorithm was determined through different comparisons which were carried out in state-of-the-art works in terms of reconstruction error, polynomial size, and computation cost. The obtained results indicate that the proposed algorithm is reliable and computes lesser coefficients when compared to the existing algorithms across wide ranges of parameter values of p and polynomial sizes N. The results also show that the improvement ratio of the computed coefficients ranges from 18.64% to 81.55% in comparison to the existing algorithms. Besides this, the proposed algorithm can generate polynomials of an order ∼8.5 times larger than those generated using state-of-the-art algorithms.
Abstract
Suffering the human because of pressure normal life of exposure to several types of heart disease as a result of due to different factors. Therefore, and in order to find out the case of a death whether or not, are to be modeled using binary logistic regression model
In this research used, one of the most important models of nonlinear regression models extensive use in the modeling of applications statistical, in terms of heart disease which is the binary logistic regression model. and then estimating the parameters of this model using the statistical estimation methods, another problem will be appears in estimating its parameters, as well as when the numbe
... Show More
In this paper, the speed control of the real DC motor is experimentally investigated using nonlinear PID neural network controller. As a simple and fast tuning algorithm, two optimization techniques are used; trial and error method and particle swarm optimization PSO algorithm in order to tune the nonlinear PID neural controller's parameters and to find best speed response of the DC motor. To save time in the real system, a Matlab simulation package is used to carry out these algorithms to tune and find the best values of the nonlinear PID parameters. Then these parameters are used in the designed real time nonlinear PID controller system based on LabVIEW package. Simulation and experimental results are compared with each other and showe
... Show MoreIn its theoretical framework, this study dealt with the subjects of high commitment management and organizational excellence, as the study came in response to the growing developments and changes in the fields of management. It includes an analysis of correlation and effect between high commitment management, which has been attracting a lot of attention recently due to the intensifying rivalry between organizations because of certain external factors like globalization and world markets liberation, and its effect in achieving organizational excellence.
The practical framework, on the other hand, dealt with the analysis of correlation and effect between the study's variables. The problem
... Show MoreThe cuneiform images need many processes in order to know their contents
and by using image enhancement to clarify the objects (symbols) founded in the
image. The Vector used for classifying the symbol called symbol structural vector
(SSV) it which is build from the information wedges in the symbol.
The experimental tests show insome numbersand various relevancy including
various drawings in online method. The results are high accuracy in this research,
and methods and algorithms programmed using a visual basic 6.0. In this research
more than one method was applied to extract information from the digital images
of cuneiform tablets, in order to identify most of signs of Sumerian cuneiform.
Many of accurate inertial guided missilc systems need to use more complex mathematical calculations and require a high speed processing to ensure the real-time opreation. This will give rise to the need of developing an effcint
Background: Sprite coding is a very effective technique for clarifying the background video object. The sprite generation is an open issue because of the foreground objects which prevent the precision of camera motion estimation and blurs the created sprite. Objective: In this paper, a quick and basic static method for sprite area detection in video data is presented. Two statistical methods are applied; the mean and standard deviation of every pixel (over all group of video frame) to determine whether the pixel is a piece of the selected static sprite range or not. A binary map array is built for demonstrating the allocated sprite (as 1) while the non-sprite (as 0) pixels valued. Likewise, holes and gaps filling strategy was utilized to re
... Show MoreSome maps of the chaotic firefly algorithm were selected to select variables for data on blood diseases and blood vessels obtained from Nasiriyah General Hospital where the data were tested and tracking the distribution of Gamma and it was concluded that a Chebyshevmap method is more efficient than a Sinusoidal map method through mean square error criterion.