Preferred Language
Articles
/
bsj-2557
Development Binary Search Algorithm
...Show More Authors

There are many methods of searching large amount of data to find one particular piece of information. Such as find name of person in record of mobile. Certain methods of organizing data make the search process more efficient the objective of these methods is to find the element with least cost (least time). Binary search algorithm is faster than sequential and other commonly used search algorithms. This research develops binary search algorithm by using new structure called Triple, structure in this structure data are represented as triple. It consists of three locations (1-Top, 2-Left, and 3-Right) Binary search algorithm divide the search interval in half, this process makes the maximum number of comparisons (Average case complexity of Search) is O(log2 n) (pronounce this "big-Oh-n" or "the order of magnitude"), if we search in a list consists of (N) elements. In this research the number of comparison is reduced to triple by using Triple structure, this process makes the maximum number of comparisons is O(log2 (n)/3+1) if we search key in list consist of (N) elements.

Crossref
View Publication Preview PDF
Quick Preview PDF
Publication Date
Fri Sep 03 2021
Journal Name
Entropy
Reliable Recurrence Algorithm for High-Order Krawtchouk Polynomials
...Show More Authors

Krawtchouk polynomials (KPs) and their moments are promising techniques for applications of information theory, coding theory, and signal processing. This is due to the special capabilities of KPs in feature extraction and classification processes. The main challenge in existing KPs recurrence algorithms is that of numerical errors, which occur during the computation of the coefficients in large polynomial sizes, particularly when the KP parameter (p) values deviate away from 0.5 to 0 and 1. To this end, this paper proposes a new recurrence relation in order to compute the coefficients of KPs in high orders. In particular, this paper discusses the development of a new algorithm and presents a new mathematical model for computing the

... Show More
View Publication
Scopus (30)
Crossref (27)
Scopus Clarivate Crossref
Publication Date
Mon Apr 24 2017
Journal Name
Ibn Al-haitham Journal For Pure And Applied Sciences
Adaptive Canny Algorithm Using Fast Otsu Multithresholding Method
...Show More Authors

   In this research, an adaptive Canny algorithm using fast Otsu multithresholding method is presented, in which fast Otsu multithresholding method is used to calculate the optimum maximum and minimum hysteresis values and used as automatic thresholding for the fourth stage of the Canny algorithm.      The new adaptive Canny algorithm and the standard Canny algorithm (manual hysteresis value) was tested on standard image (Lena) and satellite image. The results approved the validity and accuracy of the new algorithm to find the images edges for personal and satellite images as pre-step for image segmentation.  
 

View Publication Preview PDF
Publication Date
Thu Dec 31 2015
Journal Name
Al-khwarizmi Engineering Journal
Hexapod Robot Static Stability Enhancement using Genetic Algorithm
...Show More Authors

Abstract

Hexapod robot is a flexible mechanical robot with six legs. It has the ability to walk over terrain. The hexapod robot look likes the insect so it has the same gaits. These gaits are tripod, wave and ripple gaits. Hexapod robot needs to stay statically stable at all the times during each gait in order not to fall with three or more legs continuously contacts with the ground. The safety static stability walking is called (the stability margin). In this paper, the forward and inverse kinematics are derived for each hexapod’s leg in order to simulate the hexapod robot model walking using MATLAB R2010a for all gaits and the geometry in order to derive the equations of the sub-constraint workspaces for each

... Show More
View Publication Preview PDF
Publication Date
Mon Aug 01 2022
Journal Name
Bulletin Of Electrical Engineering And Informatics
Solid waste recycling and management cost optimization algorithm
...Show More Authors

Solid waste is a major issue in today's world. Which can be a contributing factor to pollution and the spread of vector-borne diseases. Because of its complicated nonlinear processes, this problem is difficult to model and optimize using traditional methods. In this study, a mathematical model was developed to optimize the cost of solid waste recycling and management. In the optimization phase, the salp swarm algorithm (SSA) is utilized to determine the level of discarded solid waste and reclaimed solid waste. An optimization technique SSA is a new method of finding the ideal solution for a mathematical relationship based on leaders and followers. It takes a lot of random solutions, as well as their outward or inward fluctuations, t

... Show More
View Publication Preview PDF
Scopus (2)
Crossref (3)
Scopus Crossref
Publication Date
Sun Jun 20 2021
Journal Name
Baghdad Science Journal
Multifactor Algorithm for Test Case Selection and Ordering
...Show More Authors

Regression testing being expensive, requires optimization notion. Typically, the optimization of test cases results in selecting a reduced set or subset of test cases or prioritizing the test cases to detect potential faults at an earlier phase. Many former studies revealed the heuristic-dependent mechanism to attain optimality while reducing or prioritizing test cases. Nevertheless, those studies were deprived of systematic procedures to manage tied test cases issue. Moreover, evolutionary algorithms such as the genetic process often help in depleting test cases, together with a concurrent decrease in computational runtime. However, when examining the fault detection capacity along with other parameters, is required, the method falls sh

... Show More
View Publication Preview PDF
Scopus (4)
Crossref (2)
Scopus Clarivate Crossref
Publication Date
Sun Dec 01 2002
Journal Name
Iraqi Journal Of Physics
An edge detection algorithm matching visual contour perception
...Show More Authors

For several applications, it is very important to have an edge detection technique matching human visual contour perception and less sensitive to noise. The edge detection algorithm describes in this paper based on the results obtained by Maximum a posteriori (MAP) and Maximum Entropy (ME) deblurring algorithms. The technique makes a trade-off between sharpening and smoothing the noisy image. One of the advantages of the described algorithm is less sensitive to noise than that given by Marr and Geuen techniques that considered to be the best edge detection algorithms in terms of matching human visual contour perception.

View Publication Preview PDF
Publication Date
Wed Sep 01 2021
Journal Name
Baghdad Science Journal
Optimum Median Filter Based on Crow Optimization Algorithm
...Show More Authors

          A novel median filter based on crow optimization algorithms (OMF) is suggested to reduce the random salt and pepper noise and improve the quality of the RGB-colored and gray images. The fundamental idea of the approach is that first, the crow optimization algorithm detects noise pixels, and that replacing them with an optimum median value depending on a criterion of maximization fitness function. Finally, the standard measure peak signal-to-noise ratio (PSNR), Structural Similarity, absolute square error and mean square error have been used to test the performance of suggested filters (original and improved median filter) used to removed noise from images. It achieves the simulation based on MATLAB R2019b and the resul

... Show More
View Publication Preview PDF
Scopus (4)
Crossref (2)
Scopus Clarivate Crossref
Publication Date
Mon Feb 01 2021
Journal Name
International Journal Of Electrical And Computer Engineering (ijece)
Features of genetic algorithm for plain text encryption
...Show More Authors

The data communication has been growing in present day. Therefore, the data encryption became very essential in secured data transmission and storage and protecting data contents from intruder and unauthorized persons. In this paper, a fast technique for text encryption depending on genetic algorithm is presented. The encryption approach is achieved by the genetic operators Crossover and mutation. The encryption proposal technique based on dividing the plain text characters into pairs, and applying the crossover operation between them, followed by the mutation operation to get the encrypted text. The experimental results show that the proposal provides an important improvement in encryption rate with comparatively high-speed Process

... Show More
View Publication
Scopus (14)
Crossref (4)
Scopus Crossref
Publication Date
Mon Oct 01 2018
Journal Name
Journal Of Economics And Administrative Sciences
Nurse Scheduling Problem Using Hybrid Simulated Annealing Algorithm
...Show More Authors

Nurse scheduling problem is one of combinatorial optimization problems and it is one of NP-Hard problems which is difficult to be solved as optimal solution. In this paper, we had created an proposed algorithm which it is hybrid simulated annealing algorithm to solve nurse scheduling problem, developed the simulated annealing algorithm and Genetic algorithm. We can note that the proposed algorithm (Hybrid simulated Annealing Algorithm(GS-h)) is the best method among other methods which it is used in this paper because it satisfied minimum average of the total cost and maximum number of Solved , Best and Optimal problems. So we can note that the ratios of the optimal solution are 77% for the proposed algorithm(GS-h), 28.75% for Si

... Show More
View Publication Preview PDF
Crossref
Publication Date
Sun Mar 04 2018
Journal Name
Iraqi Journal Of Science
Image Encryption Using DNA Encoding and RC4 Algorithm
...Show More Authors

Nowadays, the rapid development of multi-media technology and digital images transmission by the Internet leads the digital images to be exposed to several attacks in the transmission process. Therefore, protection of digital images become increasingly important.

 To this end, an image encryption method that adopts Rivest Cipher (RC4) and Deoxyribonucleic Acid (DNA) encoding to increase the secrecy and randomness of the image without affecting its quality is proposed. The Means Square Error (MSE), Peak Signal-to-Noise Ratio (PSNR), Coefficient Correlation (CC) and histogram analysis are used as an evaluation metrics to evaluate the performance of the proposed method. The results indicate that the proposed method is secure ag

... Show More
View Publication Preview PDF