Clustering is an unsupervised learning method that classified data according to similarity probabilities. DBScan as a high-quality algorithm has been introduced for clustering spatial data due to its ability to remove noise (outlier) and constructing arbitrarily shapes. However, it has a problem in determining a suitable value of Eps parameter. This paper proposes a new clustering method, termed as DBScanBAT, that it optimizes DBScan algorithm by BAT algorithm. The proposed method automatically sets the DBScan parameters (Eps) and finds the optimal value for it. The results of the proposed DBScanBAT automatically generates near original number of clusters better than DBScanPSO and original DBScan. Furthermore, the proposed method has the ability to generate high quality clusters with minimum entropy [ 0.2752, 0.4291] in TR11 and TR12 datasets.
The estimation of the parameters of linear regression is based on the usual Least Square method, as this method is based on the estimation of several basic assumptions. Therefore, the accuracy of estimating the parameters of the model depends on the validity of these hypotheses. The most successful technique was the robust estimation method which is minimizing maximum likelihood estimator (MM-estimator) that proved its efficiency in this purpose. However, the use of the model becomes unrealistic and one of these assumptions is the uniformity of the variance and the normal distribution of the error. These assumptions are not achievable in the case of studying a specific problem that may include complex data of more than one model. To
... Show MoreFuzzy C-means (FCM) is a clustering method used for collecting similar data elements within the group according to specific measurements. Tabu is a heuristic algorithm. In this paper, Probabilistic Tabu Search for FCM implemented to find a global clustering based on the minimum value of the Fuzzy objective function. The experiments designed for different networks, and cluster’s number the results show the best performance based on the comparison that is done between the values of the objective function in the case of using standard FCM and Tabu-FCM, for the average of ten runs.
Steganography is an important class of security which is widely used in computer and network security nowadays. In this research, a new proposed algorithm was introduced with a new concept of dealing with steganography as an algorithmic secret key technique similar to stream cipher cryptographic system. The proposed algorithm is a secret key system suggested to be used in communications for messages transmission steganography
Crow Search Algorithm (CSA) can be defined as one of the new swarm intelligence algorithms that has been developed lately, simulating the behavior of a crow in a storage place and the retrieval of the additional food when required. In the theory of the optimization, a crow represents a searcher, the surrounding environment represents the search space, and the random storage of food location represents a feasible solution. Amongst all the food locations, the one where the maximum amount of the food is stored is considered as the global optimum solution, and objective function represents the food amount. Through the simulation of crows’ intelligent behavior, the CSA attempts to find the optimum solutions to a variety of the proble
... Show MoreGumbel distribution was dealt with great care by researchers and statisticians. There are traditional methods to estimate two parameters of Gumbel distribution known as Maximum Likelihood, the Method of Moments and recently the method of re-sampling called (Jackknife). However, these methods suffer from some mathematical difficulties in solving them analytically. Accordingly, there are other non-traditional methods, like the principle of the nearest neighbors, used in computer science especially, artificial intelligence algorithms, including the genetic algorithm, the artificial neural network algorithm, and others that may to be classified as meta-heuristic methods. Moreover, this principle of nearest neighbors has useful statistical featu
... Show MoreRivest Cipher 4 (RC4) is an efficient stream cipher that is commonly used in internet protocols. However, there are several flaws in the key scheduling algorithm (KSA) of RC4. The contribution of this paper is to overcome some of these weaknesses by proposing a new version of KSA coined as modified KSA . In the initial state of the array is suggested to contain random values instead of the identity permutation. Moreover, the permutation of the array is modified to depend on the key value itself. The proposed performance is assessed in terms of cipher secrecy, randomness test and time under a set of experiments with variable key size and different plaintext size. The results show that the RC4 with improves the randomness and secrecy with
... Show MoreCredit risk assessment has become an important topic in financial risk administration. Fuzzy clustering analysis has been applied in credit scoring. Gustafson-Kessel (GK) algorithm has been utilised to cluster creditworthy customers as against non-creditworthy ones. A good clustering analysis implemented by good Initial Centres of clusters should be selected. To overcome this problem of Gustafson-Kessel (GK) algorithm, we proposed a modified version of Kohonen Network (KN) algorithm to select the initial centres. Utilising similar degree between points to get similarity density, and then by means of maximum density points selecting; the modified Kohonen Network method generate clustering initial centres to get more reasonable clustering res
... Show MoreThis paper presents a hybrid genetic algorithm (hGA) for optimizing the maximum likelihood function ln(L(phi(1),theta(1)))of the mixed model ARMA(1,1). The presented hybrid genetic algorithm (hGA) couples two processes: the canonical genetic algorithm (cGA) composed of three main steps: selection, local recombination and mutation, with the local search algorithm represent by steepest descent algorithm (sDA) which is defined by three basic parameters: frequency, probability, and number of local search iterations. The experimental design is based on simulating the cGA, hGA, and sDA algorithms with different values of model parameters, and sample size(n). The study contains comparison among these algorithms depending on MSE value. One can conc
... Show More