Solving problems via artificial intelligence techniques has widely prevailed in different aspects. Implementing artificial intelligence optimization algorithms for NP-hard problems is still challenging. In this manuscript, we work on implementing the Naked Mole-Rat Algorithm (NMRA) to solve the n-queens problems and overcome the challenge of applying NMRA to a discrete space set. An improvement of NMRA is applied using the aspect of local search in the Variable Neighborhood Search algorithm (VNS) with 2-opt and 3-opt. Introducing the Naked Mole Rat algorithm based on variable neighborhood search (NMRAVNS) to solve N-queens problems with different sizes. Finding the best solution or set of solutions within a plausible amount of time is the main goal of the NMRAVNS algorithm. The improvement of the proposed algorithm boosts the exploitation capability of the basic NMRA and gives a greater possibility, with the emerging search strategies, to find the global best solution. This algorithm proved successful and outperformed other algorithms and studies with a remarkable target. A detailed comparison is performed, and the data results are presented with the relevant numbers and values. NMRA and NMRAVNS comparisons are implemented and recorded. Later on, a comparison between the Meerkat Clan Algorithm, Genetic Algorithm, Particle Swarm Optimization, and NMRAVNS is tested and presented. Finally, NMRAVNS is evaluated against the examined genetic-based algorithm and listed to prove the success of the proposed algorithm. NMRAVNS outperformed previous findings and scored competitive results with a high number of queen sizes, where an average time reduction reached about 87% of other previous findings.
<p>In combinatorial testing development, the fabrication of covering arrays is the key challenge by the multiple aspects that influence it. A wide range of combinatorial problems can be solved using metaheuristic and greedy techniques. Combining the greedy technique utilizing a metaheuristic search technique like hill climbing (HC), can produce feasible results for combinatorial tests. Methods based on metaheuristics are used to deal with tuples that may be left after redundancy using greedy strategies; then the result utilization is assured to be near-optimal using a metaheuristic algorithm. As a result, the use of both greedy and HC algorithms in a single test generation system is a good candidate if constructed correctly. T
... Show MoreObjective of this work is the mixing between human biometric characteristics and unique attributes of the computer in order to protect computer networks and resources environments through the development of authentication and authorization techniques. In human biometric side has been studying the best methods and algorithms used, and the conclusion is that the fingerprint is the best, but it has some flaws. Fingerprint algorithm has been improved so that their performance can be adapted to enhance the clarity of the edge of the gully structures of pictures fingerprint, taking into account the evaluation of the direction of the nearby edges and repeat. In the side of the computer features, computer and its components like human have uniqu
... Show MoreCryptography can be thought of as a toolbox, where potential attackers gain access to various computing resources and technologies to try to compute key values. In modern cryptography, the strength of the encryption algorithm is only determined by the size of the key. Therefore, our goal is to create a strong key value that has a minimum bit length that will be useful in light encryption. Using elliptic curve cryptography (ECC) with Rubik's cube and image density, the image colors are combined and distorted, and by using the Chaotic Logistics Map and Image Density with a secret key, the Rubik's cubes for the image are encrypted, obtaining a secure image against attacks. ECC itself is a powerful algorithm that generates a pair of p
... Show MoreAbstract
Although the rapid development in reverse engineering techniques, 3D laser scanners can be considered the modern technology used to digitize the 3D objects, but some troubles may be associate this process due to the environmental noises and limitation of the used scanners. So, in the present paper a data pre-processing algorithm has been proposed to obtain the necessary geometric features and mathematical representation of scanned object from its point cloud which obtained using 3D laser scanner (Matter and Form) through isolating the noised points. The proposed algorithm based on continuous calculations of chord angle between each adjacent pair of points in point cloud. A MATLAB program has been built t
... Show MoreLet be an n-Banach space, M be a nonempty closed convex subset of , and S:M→M be a mapping that belongs to the class mapping. The purpose of this paper is to study the stability and data dependence results of a Mann iteration scheme on n-Banach space
The launch of the EU’s Eastern Partnership in 2009 intended to signal a new, elevated level of EU engagement with its Eastern neighborhood. Yet there remain several long-simmering and potentially destabilizing conflicts in the region, with which EU engagement thus far has been sporadic at best. The Union’s use of its Common Security and Defense Policy (CSDP) in the region and to help solve these disputes has been particularly ad hoc and inconsistent, wracked by inter-institutional incoherence and undermined by Member States’ inability to agree on a broad strategic vision for engagement with the area.
The three CSDP missions deployed to the region thus far have all suffered from this incoherence to various extents. In particu
... Show MoreIn this paper, The Grobner basis of the Toric Ideal for - contingency tables related with the Markov basis B introduced by Hussein S. MH, Abdulrahman H. M in 2018 is found. Also, the Grobner basis is a reduced and universal Grobner basis are shown.
In this paper, we deal with the problem of general matching of two images one of them has experienced geometrical transformations, to find the correspondence between two images. We develop the invariant moments for traditional techniques (moments of inertia) with new approach to enhance the performance for these methods. We test various projections directional moments, to extract the difference between Block Distance Moment (BDM) and evaluate their reliability. Three adaptive strategies are shown for projections directional moments, that are raster (vertical and horizontal) projection, Fan-Bean projection and new projection procedure that is the square projection method. Our paper started with the description of a new algorithm that is low
... Show MoreThis work bases on encouraging a generous and conceivable estimation for modified an algorithm for vehicle travel times on a highway from the eliminated traffic information using set aside camera image groupings. The strategy for the assessment of vehicle travel times relies upon the distinctive verification of traffic state. The particular vehicle velocities are gotten from acknowledged vehicle positions in two persistent images by working out the distance covered all through elapsed past time doing mollification between the removed traffic flow data and cultivating a plan to unequivocally predict vehicle travel times. Erbil road data base is used to recognize road locales around road segments which are projected into the commended camera
... Show More