This work aims to develop a secure lightweight cipher algorithm for constrained devices. A secure communication among constrained devices is a critical issue during the data transmission from the client to the server devices. Lightweight cipher algorithms are defined as a secure solution for constrained devices that require low computational functions and small memory. In contrast, most lightweight algorithms suffer from the trade-off between complexity and speed in order to produce robust cipher algorithm. The PRESENT cipher has been successfully experimented on as a lightweight cryptography algorithm, which transcends other ciphers in terms of its computational processing that required low complexity operations. The mathematical model of the present algorithm is simple, and the running operations required small execution time for encryption-decryption sensing data. Hence, a developed algorithm called DPRESENT was introduced to improve the complexity of the cipher text based on the PRESENT algorithm and DNA cryptography technique for developing a lightweight cipher algorithm. The NIST suite showed that the proposed algorithm tests presented high level of randomness and complexity. The execution time for the proposed algorithm was kept minimal as the current cipher algorithm. The developed algorithm is a new trend that can be applied for different lightweight cryptosystems to achieve the trade-off among complexity and speed as a robust cipher algorithm.
Most of the medical datasets suffer from missing data, due to the expense of some tests or human faults while recording these tests. This issue affects the performance of the machine learning models because the values of some features will be missing. Therefore, there is a need for a specific type of methods for imputing these missing data. In this research, the salp swarm algorithm (SSA) is used for generating and imputing the missing values in the pain in my ass (also known Pima) Indian diabetes disease (PIDD) dataset, the proposed algorithm is called (ISSA). The obtained results showed that the classification performance of three different classifiers which are support vector machine (SVM), K-nearest neighbour (KNN), and Naïve B
... Show MoreIn many areas, such as simulation, numerical analysis, computer programming, decision-making, entertainment, and coding, a random number input is required. The pseudo-random number uses its seed value. In this paper, a hybrid method for pseudo number generation is proposed using Linear Feedback Shift Registers (LFSR) and Linear Congruential Generator (LCG). The hybrid method for generating keys is proposed by merging technologies. In each method, a new large in key-space group of numbers were generated separately. Also, a higher level of secrecy is gained such that the internal numbers generated from LFSR are combined with LCG (The adoption of roots in non-linear iteration loops). LCG and LFSR are linear structures and outputs
... Show MoreFluoroscopic images are a field of medical images that depends on the quality of image for correct diagnosis; the main trouble is the de-nosing and how to keep the poise between degradation of noisy image, from one side, and edge and fine details preservation, from the other side, especially when fluoroscopic images contain black and white type noise with high density. The previous filters could usually handle low/medium black and white type noise densities, that expense edge, =fine details preservation and fail with high density of noise that corrupts the images. Therefore, this paper proposed a new Multi-Line algorithm that deals with high-corrupted image with high density of black and white type noise. The experiments achieved i
... Show MoreIn project management process, the objective is to define and develop a model for planning, scheduling, controlling, and monitoring different activities of a particular project. Time scheduling plays an important role in successful implementation of various activities and general outcome of project. In practice, various factors cause projects to suffer from time delay in accomplishing the activities. One important reason is imprecise knowledge about time duration of activities. This study addresses the problem of project scheduling in uncertain resource environments, which are defined by uncertain activity durations. The study presents a solution of the levelling and allocation problems for projects that have some uncertain ac
... Show MoreGiven the high importance of attendance for university students, upon which the possibility of keeping or losing their places in the course is based, it is essential to replace the inefficient manual method of attendance recording with a more efficient one. To handle this problem, technology must be introduced into this process. This paper aims to propose an automatic attendance system based on passive Radio Frequency Identification (RFID), fog, and cloud computing technologies (AASCF). The system has three sides. The first one, which is the Client-side; works on collecting the attendance data then sending a copy from it. The second side, which is the Server-side, works on calculating an absence ratio of all the students during the
... Show MoreA novel median filter based on crow optimization algorithms (OMF) is suggested to reduce the random salt and pepper noise and improve the quality of the RGB-colored and gray images. The fundamental idea of the approach is that first, the crow optimization algorithm detects noise pixels, and that replacing them with an optimum median value depending on a criterion of maximization fitness function. Finally, the standard measure peak signal-to-noise ratio (PSNR), Structural Similarity, absolute square error and mean square error have been used to test the performance of suggested filters (original and improved median filter) used to removed noise from images. It achieves the simulation based on MATLAB R2019b and the resul
... Show MoreTraffic management at road intersections is a complex requirement that has been an important topic of research and discussion. Solutions have been primarily focused on using vehicular ad hoc networks (VANETs). Key issues in VANETs are high mobility, restriction of road setup, frequent topology variations, failed network links, and timely communication of data, which make the routing of packets to a particular destination problematic. To address these issues, a new dependable routing algorithm is proposed, which utilizes a wireless communication system between vehicles in urban vehicular networks. This routing is position-based, known as the maximum distance on-demand routing algorithm (MDORA). It aims to find an optimal route on a hop-by-ho
... Show MoreA robust video-bitrate adaptive scheme at client-aspect plays a significant role in keeping a good quality of video streaming technology experience. Video quality affects the amount of time the video has turned off playing due to the unfilled buffer state. Therefore to maintain a video streaming continuously with smooth bandwidth fluctuation, a video buffer structure based on adapting the video bitrate is considered in this work. Initially, the video buffer structure is formulated as an optimal control-theoretic problem that combines both video bitrate and video buffer feedback signals. While protecting the video buffer occupancy from exceeding the limited operating level can provide continuous video str
... Show More<p>In combinatorial testing development, the fabrication of covering arrays is the key challenge by the multiple aspects that influence it. A wide range of combinatorial problems can be solved using metaheuristic and greedy techniques. Combining the greedy technique utilizing a metaheuristic search technique like hill climbing (HC), can produce feasible results for combinatorial tests. Methods based on metaheuristics are used to deal with tuples that may be left after redundancy using greedy strategies; then the result utilization is assured to be near-optimal using a metaheuristic algorithm. As a result, the use of both greedy and HC algorithms in a single test generation system is a good candidate if constructed correctly. T
... Show MoreBecause of the rapid development and use of the Internet as a communication media emerged to need a high level of security during data transmission and one of these ways is "Steganography". This paper reviews the Least Signification Bit steganography used for embedding text file with related image in gray-scale image. As well as we discuss the bit plane which is divided into eight different images when combination them we get the actual image. The findings of the research was the stego-image is indistinguishable to the naked eye from the original cover image when the value of bit less than four Thus we get to the goal is to cover up the existence of a connection or hidden data. The Peak to Signal Noise Ratio(PSNR) and Mean Square Error (
... Show More