This paper includes a comparison between denoising techniques by using statistical approach, principal component analysis with local pixel grouping (PCA-LPG), this procedure is iterated second time to further improve the denoising performance, and other enhancement filters were used. Like adaptive Wiener low pass-filter to a grayscale image that has been degraded by constant power additive noise, based on statistics estimated from a local neighborhood of each pixel. Performs Median filter of the input noisy image, each output pixel contains the Median value in the M-by-N neighborhood around the corresponding pixel in the input image, Gaussian low pass-filter and Order-statistic filter also be used. Experimental results shows LPG-PCA method gives better performance, especially in image fine structure preservation, compared with other general denoising algorithms.
The integer simulation and development finite impulse response (FIR) filters taking into account the possibilities of their realization on digital integer platforms are considered. The problem statement and solution of multifunctional synthesis of digital FIR filters such a problem on the basis of the numerical methods of integer nonlinear mathematical programming are given. As an several examples, the problem solution of synthesis FIR-filters with short coefficient word length has been given. The analysis of their characteristics is resulted. The paper discusses issues of modeling and synthesis of digital FIR filters with provision for the possibilities of their implementation on digital platforms with integer computation arithme
... Show MoreThe objective of the study is to demonstrate the predictive ability is better between the logistic regression model and Linear Discriminant function using the original data first and then the Home vehicles to reduce the dimensions of the variables for data and socio-economic survey of the family to the province of Baghdad in 2012 and included a sample of 615 observation with 13 variable, 12 of them is an explanatory variable and the depended variable is number of workers and the unemployed.
Was conducted to compare the two methods above and it became clear by comparing the logistic regression model best of a Linear Discriminant function written
... Show MoreIn this paper, the deterministic and the stochastic models are proposed to study the interaction of the Coronavirus (COVID-19) with host cells inside the human body. In the deterministic model, the value of the basic reproduction number determines the persistence or extinction of the COVID-19. If , one infected cell will transmit the virus to less than one cell, as a result, the person carrying the Coronavirus will get rid of the disease .If the infected cell will be able to infect all cells that contain ACE receptors. The stochastic model proves that if are sufficiently large then maybe give us ultimate disease extinction although , and this facts also proved by computer simulation.
In this paper, we will illustrate a gamma regression model assuming that the dependent variable (Y) is a gamma distribution and that it's mean ( ) is related through a linear predictor with link function which is identity link function g(μ) = μ. It also contains the shape parameter which is not constant and depends on the linear predictor and with link function which is the log link and we will estimate the parameters of gamma regression by using two estimation methods which are The Maximum Likelihood and the Bayesian and a comparison between these methods by using the standard comparison of average squares of error (MSE), where the two methods were applied to real da
... Show MoreIn this paper was discussed the process of compounding two distributions using new compounding procedure which is connect a number of life time distributions ( continuous distribution ) where is the number of these distributions represent random variable distributed according to one of the discrete random distributions . Based on this procedure have been compounding zero – truncated poisson distribution with weibell distribution to produce new life time distribution having three parameter , Advantage of that failure rate function having many cases ( increasing , dicreasing , unimodal , bathtube) , and study the resulting distribution properties such as : expectation , variance , comulative function , reliability function and fa
... Show MoreTwilight is that light appear on the horizon before sunrise and after sunset, Astronomically it is known that sunrise and sunset are effected by high above sea level, but the effect of high above sea level on the time of astronomical twilight still not decided and controversy among astronomers, in This research we studies the effect of high above sea level on the time of astronomical twilight, through adding the equation correct high above sea level to equation computation of twilight and then calculate of changing in the time of twilight for different highest (0-10000) meters above sea level , and the ratio of increase for time with high between (15.45-20.5) minutes. It was found that there was an increase in the time of the twilight along
... Show MoreThe proliferation of many editing programs based on artificial intelligence techniques has contributed to the emergence of deepfake technology. Deepfakes are committed to fabricating and falsifying facts by making a person do actions or say words that he never did or said. So that developing an algorithm for deepfakes detection is very important to discriminate real from fake media. Convolutional neural networks (CNNs) are among the most complex classifiers, but choosing the nature of the data fed to these networks is extremely important. For this reason, we capture fine texture details of input data frames using 16 Gabor filters indifferent directions and then feed them to a binary CNN classifier instead of using the red-green-blue
... Show More