Excessive skewness which occurs sometimes in the data is represented as an obstacle against normal distribution. So, recent studies have witnessed activity in studying the skew-normal distribution (SND) that matches the skewness data which is regarded as a special case of the normal distribution with additional skewness parameter (α), which gives more flexibility to the normal distribution. When estimating the parameters of (SND), we face the problem of the non-linear equation and by using the method of Maximum Likelihood estimation (ML) their solutions will be inaccurate and unreliable. To solve this problem, two methods can be used that are: the genetic algorithm (GA) and the iterative reweighting algorithm (IR) based on the Maximum Likelihood method. Monte Carlo simulation was used with different skewness levels and sample sizes, and the superiority of the results was compared. It was concluded that (SND) model estimation using (GA) is the best when the samples sizes are small and medium, while large samples indicate that the (IR) algorithm is the best. The study was also done using real data to find the parameter estimation and a comparison between the superiority of the results based on (AIC, BIC, Mse and Def) criteria.
In this paper, some estimators for the unknown shape parameters and reliability function of Basic Gompertz distribution were obtained, such as Maximum likelihood estimator and some Bayesian estimators under Squared log error loss function by using Gamma and Jefferys priors. Monte-Carlo simulation was conducted to compare the performance of all estimates of the shape parameter and Reliability function, based on mean squared errors (MSE) and integrated mean squared errors (IMSE's), respectively. Finally, the discussion is provided to illustrate the results that are summarized in tables.
The main aim of this paper is to study how the different estimators of the two unknown parameters (shape and scale parameter) of a generalized exponential distribution behave for different sample sizes and for different parameter values. In particular,
. Maximum Likelihood, Percentile and Ordinary Least Square estimators had been implemented for different sample sizes (small, medium, and large) and assumed several contrasts initial values for the two parameters. Two indicators of performance Mean Square Error and Mean Percentile Error were used and the comparisons were carried out between different methods of estimation by using monte carlo simulation technique .. It was obse
... Show MoreThis paper uses Artificial Intelligence (AI) based algorithm analysis to classify breast cancer Deoxyribonucleic (DNA). Main idea is to focus on application of machine and deep learning techniques. Furthermore, a genetic algorithm is used to diagnose gene expression to reduce the number of misclassified cancers. After patients' genetic data are entered, processing operations that require filling the missing values using different techniques are used. The best data for the classification process are chosen by combining each technique using the genetic algorithm and comparing them in terms of accuracy.
Artificial fish swarm algorithm (AFSA) is one of the critical swarm intelligent algorithms. In this
paper, the authors decide to enhance AFSA via diversity operators (AFSA-DO). The diversity operators will
be producing more diverse solutions for AFSA to obtain reasonable resolutions. AFSA-DO has been used to
solve flexible job shop scheduling problems (FJSSP). However, the FJSSP is a significant problem in the
domain of optimization and operation research. Several research papers dealt with methods of solving this
issue, including forms of intelligence of the swarms. In this paper, a set of FJSSP target samples are tested
employing the improved algorithm to confirm its effectiveness and evaluate its ex
This work presents an approach for the applying Triple DES (TRIPLE DES) based on using genetic algorithm by adding intelligent feature for TRIPLE DES with N round for genetic algorithm. Encapsulated cipher file with special program which send an acknowledgment to a sender to know who decipher or broken to crash it , Thus it is considered as the initial step to improve privacy. The outcome for proposed system gives a good indication that it is a promising system compared with other type of cipher system.
Variable selection is an essential and necessary task in the statistical modeling field. Several studies have triedto develop and standardize the process of variable selection, but it isdifficultto do so. The first question a researcher needs to ask himself/herself what are the most significant variables that should be used to describe a given dataset’s response. In thispaper, a new method for variable selection using Gibbs sampler techniqueshas beendeveloped.First, the model is defined, and the posterior distributions for all the parameters are derived.The new variable selection methodis tested usingfour simulation datasets. The new approachiscompared with some existingtechniques: Ordinary Least Squared (OLS), Least Absolute Shrinkage
... Show MoreThe aim of this research is to construct a three-dimensional maritime transport model to transport nonhomogeneous goods (k) and different transport modes (v) from their sources (i) to their destinations (j), while limiting the optimum quantities v ijk x to be transported at the lowest possible cost v ijk c and time v ijk t using the heuristic algorithm, Transport problems have been widely studied in computer science and process research and are one of the main problems of transport problems that are usually used to reduce the cost or times of transport of goods with a number of sources and a number of destinations and by means of transport to meet the conditions of supply and demand. Transport models are a key tool in logistics an
... Show MoreThe modern systems that have been based upon the hash function are more suitable compared to the conventional systems; however, the complicated algorithms for the generation of the invertible functions have a high level of time consumption. With the use of the GAs, the key strength is enhanced, which results in ultimately making the entire algorithm sufficient. Initially, the process of the key generation is performed by using the results of n-queen problem that is solved by the genetic algorithm, with the use of a random number generator and through the application of the GA operations. Ultimately, the encryption of the data is performed with the use of the Modified Reverse Encryption Algorithm (MREA). It was noticed that the
... Show MoreNowadays, the power plant is changing the power industry from a centralized and vertically integrated form into regional, competitive and functionally separate units. This is done with the future aims of increasing efficiency by better management and better employment of existing equipment and lower price of electricity to all types of customers while retaining a reliable system. This research is aimed to solve the optimal power flow (OPF) problem. The OPF is used to minimize the total generations fuel cost function. Optimal power flow may be single objective or multi objective function. In this thesis, an attempt is made to minimize the objective function with keeping the voltages magnitudes of all load buses, real outp
... Show MoreCrow Search Algorithm (CSA) can be defined as one of the new swarm intelligence algorithms that has been developed lately, simulating the behavior of a crow in a storage place and the retrieval of the additional food when required. In the theory of the optimization, a crow represents a searcher, the surrounding environment represents the search space, and the random storage of food location represents a feasible solution. Amongst all the food locations, the one where the maximum amount of the food is stored is considered as the global optimum solution, and objective function represents the food amount. Through the simulation of crows’ intelligent behavior, the CSA attempts to find the optimum solutions to a variety of the proble
... Show More