Gumbel distribution was dealt with great care by researchers and statisticians. There are traditional methods to estimate two parameters of Gumbel distribution known as Maximum Likelihood, the Method of Moments and recently the method of re-sampling called (Jackknife). However, these methods suffer from some mathematical difficulties in solving them analytically. Accordingly, there are other non-traditional methods, like the principle of the nearest neighbors, used in computer science especially, artificial intelligence algorithms, including the genetic algorithm, the artificial neural network algorithm, and others that may to be classified as meta-heuristic methods. Moreover, this principle of nearest neighbors has useful statistical features. The objective of this paper is thus to propose a new algorithm where it allows getting the estimation of the parameters of Gumbel probability distribution directly. Furthermore, it overcomes the mathematical difficulties in this matter without need to the derivative of the likelihood function. Taking simulation approach under consideration as empirical experiments where a hybrid method performs optimization of these three traditional methods. In this regard, comparisons have been done between the new proposed method and each pair of the traditional methods mentioned above by efficiency criterion Root of Mean Squared Error (RMSE). As a result, (36) experiments of different combinations of initial values of two parameters (λ: shift parameter and θ: scale parameter) in three values that take four different sample sizes for each experiment. To conclude, the proposed algorithm showed its superiority in all simulation combinations associated with all sample sizes for the two parameters (λ and θ). In addition, the method of Moments was the best in estimating the shift parameter (λ) and the method of Maximum Likelihood was in estimating the scale parameter (θ).
In this research, the focus was placed on estimating the parameters of the Hypoexponential distribution function using the maximum likelihood method and genetic algorithm. More than one standard, including MSE, has been adopted for comparison by Using the simulation method
There are many studies dealt with handoff management in mobile communication systems and some of these studies presented handoff schemes to manage this important process in cellular network. All previous schemes used relative signal strength (RSS) measurements. In this work, a new proposed handoff scheme had been presented depending not only on the RSS measurements but also used the threshold distance and neighboring BSS power margins in order to improve the handoff management process. We submitted here a threshold RSS as a condition to make a handoff when a mobile station moves from one cell to another this at first, then we submitted also a specified margin between the current received signal and the ongoing BS's received signal must be s
... Show Moreإن استخدام النظم الالكترونية في القطاع المصرفي وبالخصوص نظام مقاصة الصكوك الالكترونية (ACH) في عمليات التحويل الالكتروني للاموال بين المصارف تتضمن تحويلات مالية عالية القيمة بين المصارف المشاركة بهذا النظام, وان اي خلل قد يحدث بالنظام يؤدي الى حالات تلاعب في مقاصة الصكوك الالكترونية في المصارف المشاركة وبالتالي حدوث عملية اختلاس, ومن هذا المنطلق تبرز مشكلة البحث في اهمية توافر برنامج تدقيق مقترح ياخ
... Show MoreThe problem motivation of this work deals with how to control the network overhead and reduce the network latency that may cause many unwanted loops resulting from using standard routing. This work proposes three different wireless routing protocols which they are originally using some advantages for famous wireless ad-hoc routing protocols such as dynamic source routing (DSR), optimized link state routing (OLSR), destination sequenced distance vector (DSDV) and zone routing protocol (ZRP). The first proposed routing protocol is presented an enhanced destination sequenced distance vector (E-DSDV) routing protocol, while the second proposed routing protocol is designed based on using the advantages of DSDV and ZRP and we named it as
... Show MoreIn this paper, we are mainly concerned with estimating cascade reliability model (2+1) based on inverted exponential distribution and comparing among the estimation methods that are used . The maximum likelihood estimator and uniformly minimum variance unbiased estimators are used to get of the strengths and the stress ;k=1,2,3 respectively then, by using the unbiased estimators, we propose Preliminary test single stage shrinkage (PTSSS) estimator when a prior knowledge is available for the scale parameter as initial value due past experiences . The Mean Squared Error [MSE] for the proposed estimator is derived to compare among the methods. Numerical results about conduct of the considered
... Show MoreIn this work, an estimation of the key rate of measurement-device-independent quantum key distribution (MDI-QKD) protocol in free space was performed. The examined free space links included satellite-earth downlink, uplink and intersatellite link. Various attenuation effects were considered such as diffraction, atmosphere, turbulence and the efficiency of the detection system. Two cases were tested: asymptotic case with infinite number of decoy states and one-decoy state case. The estimated key rate showed the possibility of applying MDI-QKD in earth-satellite and intersatellite links, offering longer single link distance to be covered.
Abstract
In this study, we compare between the autoregressive approximations (Yule-Walker equations, Least Squares , Least Squares ( forward- backword ) and Burg’s (Geometric and Harmonic ) methods, to determine the optimal approximation to the time series generated from the first - order moving Average non-invertible process, and fractionally - integrated noise process, with several values for d (d=0.15,0.25,0.35,0.45) for different sample sizes (small,median,large)for two processes . We depend on figure of merit function which proposed by author Shibata in 1980, to determine the theoretical optimal order according to min
... Show MoreA new approach for baud time (or baud rate) estimation of a random binary signal is presented. This approach utilizes the spectrum of the signal after nonlinear processing in a way that the estimation error can be reduced by simply increasing the number of the processed samples instead of increasing the sampling rate. The spectrum of the new signal is shown to give an accurate estimate about the baud time when there is no apriory information or any restricting preassumptions. The performance of the estimator for random binary square waves perturbed by white Gaussian noise and ISI is evaluated and compared with that of the conventional estimator of the zero crossing detector.
In this paper, we introduce three robust fuzzy estimators of a location parameter based on Buckley’s approach, in the presence of outliers. These estimates were compared using the variance of fuzzy numbers criterion, all these estimates were best of Buckley’s estimate. of these, the fuzzy median was the best in the case of small and medium sample size, and in large sample size, the fuzzy trimmed mean was the best.
This paper concerned with estimation reliability ( for K components parallel system of the stress-strength model with non-identical components which is subjected to a common stress, when the stress and strength follow the Generalized Exponential Distribution (GED) with unknown shape parameter α and the known scale parameter θ (θ=1) to be common. Different shrinkage estimation methods will be considered to estimate  depending on maximum likelihood estimator and prior estimates based on simulation using mean squared error (MSE) criteria. The study approved that the shrinkage estimation using shrinkage weight function was the best.