Preferred Language
Articles
/
jeasiq-2148
Using Iterative Reweighting Algorithm and Genetic Algorithm to Calculate The Estimation of The Parameters Of The Maximum Likelihood of The Skew Normal Distribution

Excessive skewness which occurs sometimes in the data is represented as an obstacle against normal distribution. So, recent studies have witnessed activity in studying the skew-normal distribution (SND) that matches the skewness data which is regarded as a special case of the normal distribution with additional skewness parameter (α), which gives more flexibility to the normal distribution. When estimating the parameters of (SND), we face the problem of the non-linear equation and by using the method of Maximum Likelihood estimation (ML) their solutions will be inaccurate and unreliable. To solve this problem, two methods can be used that are: the genetic algorithm (GA) and the iterative reweighting algorithm (IR) based on the Maximum Likelihood method. Monte Carlo simulation was used with different skewness levels and sample sizes, and the superiority of the results was compared. It was concluded that (SND) model estimation using (GA) is the best when the samples sizes are small and medium, while large samples indicate that the (IR) algorithm is the best. The study was also done using real data to find the parameter estimation and a comparison between the superiority of the results based on (AIC, BIC, Mse and Def) criteria.

Crossref
View Publication Preview PDF
Quick Preview PDF
Publication Date
Wed Jul 01 2015
Journal Name
Journal Of Engineering
Optimal Location of Static Synchronous Compensator (STATCOM) for IEEE 5-Bus Standard System Using Genetic Algorithm

Heuristic approaches are traditionally applied to find the optimal size and optimal location of Flexible AC Transmission Systems (FACTS) devices in power systems. Genetic Algorithm (GA) technique has been applied to solve power engineering optimization problems giving better results than classical methods. This paper shows the application of GA for optimal sizing and allocation of a Static Compensator (STATCOM) in a power system. STATCOM devices used to increase transmission systems capacity and enhance voltage stability by regulate the voltages at its terminal by controlling the amount of reactive power injected into or absorbed from the power system. IEEE 5-bus standard system is used as an example to illustrate the te

... Show More
View Publication Preview PDF
Publication Date
Fri Feb 01 2019
Journal Name
Journal Of Economics And Administrative Sciences
A Comparison of Parameters Estimation Methods for the Negative Binomial Regression Model under Multicollinearity Problem by Using Simulation

This study discussed a biased estimator of the Negative Binomial Regression model known as (Liu Estimator), This estimate was used to reduce variance and overcome the problem Multicollinearity between explanatory variables, Some estimates were used such as Ridge Regression and Maximum Likelihood Estimators, This research aims at the theoretical comparisons between the new estimator (Liu Estimator) and the estimators

... Show More
Crossref
View Publication Preview PDF
Publication Date
Fri Dec 01 2017
Journal Name
Journal Of Economics And Administrative Sciences
A comparison between Bayesian Method and Full Maximum Likelihood to estimate Poisson regression model hierarchy and its application to the maternal deaths in Baghdad

Abstract:

 This research aims to compare Bayesian Method and Full Maximum Likelihood to estimate hierarchical Poisson regression model.

The comparison was done by  simulation  using different sample sizes (n = 30, 60, 120) and different Frequencies (r = 1000, 5000) for the experiments as was the adoption of the  Mean Square Error to compare the preference estimation methods and then choose the best way to appreciate model and concluded that hierarchical Poisson regression model that has been appreciated Full Maximum Likelihood Full Maximum Likelihood  with sample size  (n = 30) is the best to represent the maternal mortality data after it has been reliance value param

... Show More
Crossref
View Publication Preview PDF
Publication Date
Sat Jul 31 2021
Journal Name
Iraqi Journal Of Science
A Decision Tree-Aware Genetic Algorithm for Botnet Detection

     In this paper, the botnet detection problem is defined as a feature selection problem and the genetic algorithm (GA) is used to search for the best significant combination of features from the entire search space of set of features. Furthermore, the Decision Tree (DT) classifier is used as an objective function to direct the ability of the proposed GA to locate the combination of features that can correctly classify the activities into normal traffics and botnet attacks. Two datasets  namely the UNSW-NB15 and the Canadian Institute for Cybersecurity Intrusion Detection System 2017 (CICIDS2017), are used as evaluation datasets. The results reveal that the proposed DT-aware GA can effectively find the relevant

... Show More
Scopus (3)
Crossref (1)
Scopus Crossref
View Publication Preview PDF
Publication Date
Sun Jan 01 2023
Journal Name
Aip Conference Proceedings
Using the artificial TABU algorithm to estimate the semi-parametric regression function with measurement errors

Artificial Intelligence Algorithms have been used in recent years in many scientific fields. We suggest employing artificial TABU algorithm to find the best estimate of the semi-parametric regression function with measurement errors in the explanatory variables and the dependent variable, where measurement errors appear frequently in fields such as sport, chemistry, biological sciences, medicine, and epidemiological studies, rather than an exact measurement.

Scopus Crossref
Publication Date
Wed Oct 17 2018
Journal Name
Journal Of Economics And Administrative Sciences
New Robust Estimation in Compound Exponential Weibull-Poisson Distribution for both contaminated and non-contaminated Data

Abstract

The research Compared two methods for estimating fourparametersof the compound exponential Weibull - Poisson distribution which are the maximum likelihood method and the Downhill Simplex algorithm. Depending on two data cases, the first one assumed the original data (Non-polluting), while the second one assumeddata contamination. Simulation experimentswere conducted for different sample sizes and initial values of parameters and under different levels of contamination. Downhill Simplex algorithm was found to be the best method for in the estimation of the parameters, the probability function and the reliability function of the compound distribution in cases of natural and contaminateddata.

 

... Show More
Crossref
View Publication Preview PDF
Publication Date
Wed Mar 16 2022
Journal Name
2022 Muthanna International Conference On Engineering Science And Technology (micest)
Scopus (2)
Crossref (1)
Scopus Crossref
View Publication
Publication Date
Wed Aug 31 2022
Journal Name
Iraqi Journal Of Science
Generating Streams of Random Key Based on Image Chaos and Genetic Algorithm

    Today the Genetic Algorithm (GA) tops all the standard algorithms in solving complex nonlinear equations based on the laws of nature. However, permute convergence is considered one of the most significant drawbacks of GA, which is known as increasing the number of iterations needed to achieve a global optimum. To address this shortcoming, this paper proposes a new GA based on chaotic systems. In GA processes, we use the logistic map and the Linear Feedback Shift Register (LFSR) to generate chaotic values to use instead of each step requiring random values. The Chaos Genetic Algorithm (CGA) avoids local convergence more frequently than the traditional GA due to its diversity. The concept is using chaotic sequences with LFSR to gene

... Show More
Scopus (1)
Scopus Crossref
View Publication Preview PDF
Publication Date
Sun Feb 25 2024
Journal Name
Baghdad Science Journal
Natural Language Processing For Requirement Elicitation In University Using Kmeans And Meanshift Algorithm

 Data Driven Requirement Engineering (DDRE) represents a vision for a shift from the static traditional methods of doing requirements engineering to dynamic data-driven user-centered methods. Data available and the increasingly complex requirements of system software whose functions can adapt to changing needs to gain the trust of its users, an approach is needed in a continuous software engineering process. This need drives the emergence of new challenges in the discipline of requirements engineering to meet the required changes. The problem in this study was the method in data discrepancies which resulted in the needs elicitation process being hampered and in the end software development found discrepancies and could not meet the need

... Show More
Scopus (1)
Scopus Crossref
View Publication Preview PDF
Publication Date
Sun Dec 07 2014
Journal Name
Baghdad Science Journal
Baysian and NonBaysian Methods to Estimate the two parameters of Logistic Distribution

In this paper ,the problem of point estimation for the two parameters of logistic distribution has been investigated using simulation technique. The rank sampling set estimator method which is one of the Non_Baysian procedure and Lindley approximation estimator method which is one of the Baysian method were used to estimate the parameters of logistic distribution. Comparing between these two mentioned methods by employing mean square error measure and mean absolute percentage error measure .At last simulation technique used to generate many number of samples sizes to compare between these methods.

Crossref
View Publication Preview PDF