Preferred Language
Articles
/
bsj-2557
Development Binary Search Algorithm
...Show More Authors

There are many methods of searching large amount of data to find one particular piece of information. Such as find name of person in record of mobile. Certain methods of organizing data make the search process more efficient the objective of these methods is to find the element with least cost (least time). Binary search algorithm is faster than sequential and other commonly used search algorithms. This research develops binary search algorithm by using new structure called Triple, structure in this structure data are represented as triple. It consists of three locations (1-Top, 2-Left, and 3-Right) Binary search algorithm divide the search interval in half, this process makes the maximum number of comparisons (Average case complexity of Search) is O(log2 n) (pronounce this "big-Oh-n" or "the order of magnitude"), if we search in a list consists of (N) elements. In this research the number of comparison is reduced to triple by using Triple structure, this process makes the maximum number of comparisons is O(log2 (n)/3+1) if we search key in list consist of (N) elements.

Crossref
View Publication Preview PDF
Quick Preview PDF
Publication Date
Sun Nov 01 2020
Journal Name
2020 2nd Annual International Conference On Information And Sciences (aicis)
An Enhanced Multi-Objective Evolutionary Algorithm with Decomposition for Signed Community Detection Problem
...Show More Authors

View Publication
Scopus (3)
Crossref (1)
Scopus Crossref
Publication Date
Sun Jan 01 2017
Journal Name
Iraqi Journal Of Science
Strong Triple Data Encryption Standard Algorithm using Nth Degree Truncated Polynomial Ring Unit
...Show More Authors

Cryptography is the process of transforming message to avoid an unauthorized access of data. One of the main problems and an important part in cryptography with secret key algorithms is key. For higher level of secure communication key plays an important role. For increasing the level of security in any communication, both parties must have a copy of the secret key which, unfortunately, is not that easy to achieve. Triple Data Encryption Standard algorithm is weak due to its weak key generation, so that key must be reconfigured to make this algorithm more secure, effective, and strong. Encryption key enhances the Triple Data Encryption Standard algorithm securities. This paper proposed a combination of two efficient encryption algorithms to

... Show More
Publication Date
Wed Aug 30 2023
Journal Name
Iraqi Journal Of Science
Community Detection in Modular Complex Networks Using an Improved Particle Swarm Optimization Algorithm
...Show More Authors

     Community detection is an important and interesting topic for better understanding and analyzing complex network structures. Detecting hidden partitions in complex networks is proven to be an NP-hard problem that may not be accurately resolved using traditional methods. So it is solved using evolutionary computation methods and modeled in the literature as an optimization problem.  In recent years, many researchers have directed their research efforts toward addressing the problem of community structure detection by developing different algorithms and making use of single-objective optimization methods. In this study, we have continued that research line by improving the Particle Swarm Optimization (PSO) algorithm using a local

... Show More
View Publication Preview PDF
Scopus (1)
Scopus Crossref
Publication Date
Sun Jan 30 2022
Journal Name
Iraqi Journal Of Science
Modified Blowfish Algorithm for Image Encryption using Multi Keys based on five Sboxes
...Show More Authors

In this paper, a new modification was proposed to enhance the security level in the Blowfish algorithm by increasing the difficulty of cracking the original message which will lead to be safe against unauthorized attack. This algorithm is a symmetric variable-length key, 64-bit block cipher and it is implemented using gray scale images of different sizes. Instead of using a single key in cipher operation, another key (KEY2) of one byte length was used in the proposed algorithm which has taken place in the Feistel function in the first round both in encryption and decryption processes. In addition, the proposed modified Blowfish algorithm uses five Sboxes instead of four; the additional key (KEY2) is selected randomly from additional Sbox

... Show More
View Publication Preview PDF
Publication Date
Wed Jun 01 2022
Journal Name
Baghdad Science Journal
Variable Selection Using aModified Gibbs Sampler Algorithm with Application on Rock Strength Dataset
...Show More Authors

Variable selection is an essential and necessary task in the statistical modeling field. Several studies have triedto develop and standardize the process of variable selection, but it isdifficultto do so. The first question a researcher needs to ask himself/herself what are the most significant variables that should be used to describe a given dataset’s response. In thispaper, a new method for variable selection using Gibbs sampler techniqueshas beendeveloped.First, the model is defined, and the posterior distributions for all the parameters are derived.The new variable selection methodis tested usingfour simulation datasets. The new approachiscompared with some existingtechniques: Ordinary Least Squared (OLS), Least Absolute Shrinkage

... Show More
View Publication Preview PDF
Scopus (3)
Crossref (1)
Scopus Clarivate Crossref
Publication Date
Sun Mar 08 2015
Journal Name
All Days
Distribution of New Horizontal Wells by the Use of Artificial Neural Network Algorithm
...Show More Authors
Abstract<p>It is an established fact that substantial amounts of oil usually remain in a reservoir after primary and secondary processes. Therefore; there is an ongoing effort to sweep that remaining oil. Field optimization includes many techniques. Horizontal wells are one of the most motivating factors for field optimization. The selection of new horizontal wells must be accompanied with the right selection of the well locations. However, modeling horizontal well locations by a trial and error method is a time consuming method. Therefore; a method of Artificial Neural Network (ANN) has been employed which helps to predict the optimum performance via proposed new wells locations by incorporatin</p> ... Show More
View Publication
Scopus (2)
Scopus Crossref
Publication Date
Mon Jan 01 2024
Journal Name
Aip Conference Proceedings
Modeling and analysis of thermal contrast based on LST algorithm for Baghdad city
...Show More Authors

View Publication
Scopus Crossref
Publication Date
Tue May 01 2012
Journal Name
2012 Second International Conference On Digital Information And Communication Technology And It's Applications (dictap)
The compact Genetic Algorithm for likelihood estimator of first order moving average model
...Show More Authors

Recently Genetic Algorithms (GAs) have frequently been used for optimizing the solution of estimation problems. One of the main advantages of using these techniques is that they require no knowledge or gradient information about the response surface. The poor behavior of genetic algorithms in some problems, sometimes attributed to design operators, has led to the development of other types of algorithms. One such class of these algorithms is compact Genetic Algorithm (cGA), it dramatically reduces the number of bits reqyuired to store the poulation and has a faster convergence speed. In this paper compact Genetic Algorithm is used to optimize the maximum likelihood estimator of the first order moving avergae model MA(1). Simulation results

... Show More
View Publication
Scopus (5)
Scopus Crossref
Publication Date
Mon Dec 01 2014
Journal Name
2014 Ieee Symposium On Differential Evolution (sde)
Comparative analysis of a modified differential evolution algorithm based on bacterial mutation scheme
...Show More Authors

A new modified differential evolution algorithm DE-BEA, is proposed to improve the reliability of the standard DE/current-to-rand/1/bin by implementing a new mutation scheme inspired by the bacterial evolutionary algorithm (BEA). The crossover and the selection schemes of the DE method are also modified to fit the new DE-BEA mechanism. The new scheme diversifies the population by applying to all the individuals a segment based scheme that generates multiple copies (clones) from each individual one-by-one and applies the BEA segment-wise mechanism. These new steps are embedded in the DE/current-to-rand/bin scheme. The performance of the new algorithm has been compared with several DE variants over eighteen benchmark functions including sever

... Show More
View Publication
Scopus (3)
Crossref (2)
Scopus Crossref
Publication Date
Sun Apr 30 2023
Journal Name
Iraqi Journal Of Science
An Evolutionary Algorithm with Gene Ontology-Aware Crossover Operator for Protein Complex Detection
...Show More Authors

     Evolutionary algorithms (EAs), as global search methods, are proved to be more robust than their counterpart local heuristics for detecting protein complexes in protein-protein interaction (PPI) networks. Typically, the source of robustness of these EAs comes from their components and parameters. These components are solution representation, selection, crossover, and mutation. Unfortunately, almost all EA based complex detection methods suggested in the literature were designed with only canonical or traditional components. Further, topological structure of the protein network is the main information that is used in the design of almost all such components. The main contribution of this paper is to formulate a more robust E

... Show More
Scopus (1)
Scopus Crossref