Establishing coverage of the target sensing field and extending the network’s lifetime, together known as Coverage-lifetime is the key issue in wireless sensor networks (WSNs). Recent studies realize the important role of nature-inspired algorithms in handling coverage-lifetime problem with different optimization aspects. One of the main formulations is to define coverage-lifetime problem as a disjoint set covers problem. In this paper, we propose an evolutionary algorithm for solving coverage-lifetime problem as a disjoint set covers function. The main interest in this paper is to reflect both models of sensing: Boolean and probabilistic. Moreover, a heuristic operator is proposed as a local refinement operator to improve the quality of the solutions provided by the evolutionary algorithm. Simulation results show the necessity to inject a heuristic operator within the mechanism of evolutionary algorithm to improve its performance. Additionally, the results show performance difference while adopting the two types of sensing models.
Corona Virus Disease-2019 (COVID-19) is a novel virus belongs to the corona virus's family. It spreads very quickly and causes many deaths around the world. The early diagnosis of the disease can help in providing the proper therapy and saving the humans' life. However, it founded that the diagnosis of chest radiography can give an indicator of coronavirus. Thus, a Corner-based Weber Local Descriptor (CWLD) for COVID-19 diagnostics based on chest X-Ray image analysis is presented in this article. The histogram of Weber differential excitation and gradient orientation of the local regions surrounding points of interest are proposed to represent the patterns of the chest X-Ray image. Support Vector Machine (SVM) and Deep Belief Network (DBN)
... Show MoreThe aim of this study was to identify the depth of the mouth and its shape in some local fish belonging to the Cyprinidae family, and the extent to which the depth of the mouth is related to the way of feeding and the nature of food as well as the feeding habits of those species collected specifically from the Tigris River, the results showed a relationship of depth oral cavity with head length was highly significant at (P < 0.01) for all studied species. Also, there was a highly significant relationship between the height of the pharyngeal tooth-bearing bone and the depth of the oral cavity for fish of this local family.
Survival analysis is widely applied in data describing for the life time of item until the occurrence of an event of interest such as death or another event of understudy . The purpose of this paper is to use the dynamic approach in the deep learning neural network method, where in this method a dynamic neural network that suits the nature of discrete survival data and time varying effect. This neural network is based on the Levenberg-Marquardt (L-M) algorithm in training, and the method is called Proposed Dynamic Artificial Neural Network (PDANN). Then a comparison was made with another method that depends entirely on the Bayes methodology is called Maximum A Posterior (MAP) method. This method was carried out using numerical algorithms re
... Show MoreBecause of the rapid development and use of the Internet as a communication media emerged to need a high level of security during data transmission and one of these ways is "Steganography". This paper reviews the Least Signification Bit steganography used for embedding text file with related image in gray-scale image. As well as we discuss the bit plane which is divided into eight different images when combination them we get the actual image. The findings of the research was the stego-image is indistinguishable to the naked eye from the original cover image when the value of bit less than four Thus we get to the goal is to cover up the existence of a connection or hidden data. The Peak to Signal Noise Ratio(PSNR) and Mean Square Error (
... Show MoreCryptography can be thought of as a toolbox, where potential attackers gain access to various computing resources and technologies to try to compute key values. In modern cryptography, the strength of the encryption algorithm is only determined by the size of the key. Therefore, our goal is to create a strong key value that has a minimum bit length that will be useful in light encryption. Using elliptic curve cryptography (ECC) with Rubik's cube and image density, the image colors are combined and distorted, and by using the Chaotic Logistics Map and Image Density with a secret key, the Rubik's cubes for the image are encrypted, obtaining a secure image against attacks. ECC itself is a powerful algorithm that generates a pair of p
... Show More<p>In combinatorial testing development, the fabrication of covering arrays is the key challenge by the multiple aspects that influence it. A wide range of combinatorial problems can be solved using metaheuristic and greedy techniques. Combining the greedy technique utilizing a metaheuristic search technique like hill climbing (HC), can produce feasible results for combinatorial tests. Methods based on metaheuristics are used to deal with tuples that may be left after redundancy using greedy strategies; then the result utilization is assured to be near-optimal using a metaheuristic algorithm. As a result, the use of both greedy and HC algorithms in a single test generation system is a good candidate if constructed correctly. T
... Show MoreThis paper include the problem of segmenting an image into regions represent (objects), segment this object by define boundary between two regions using a connected component labeling. Then develop an efficient segmentation algorithm based on this method, to apply the algorithm to image segmentation using different kinds of images, this algorithm consist four steps at the first step convert the image gray level the are applied on the image, these images then in the second step convert to binary image, edge detection using Canny edge detection in third Are applie the final step is images. Best segmentation rates are (90%) obtained when using the developed algorithm compared with (77%) which are obtained using (ccl) before enhancement.
Abstract
Although the rapid development in reverse engineering techniques, 3D laser scanners can be considered the modern technology used to digitize the 3D objects, but some troubles may be associate this process due to the environmental noises and limitation of the used scanners. So, in the present paper a data pre-processing algorithm has been proposed to obtain the necessary geometric features and mathematical representation of scanned object from its point cloud which obtained using 3D laser scanner (Matter and Form) through isolating the noised points. The proposed algorithm based on continuous calculations of chord angle between each adjacent pair of points in point cloud. A MATLAB program has been built t
... Show More