Objective of this work is the mixing between human biometric characteristics and unique attributes of the computer in order to protect computer networks and resources environments through the development of authentication and authorization techniques. In human biometric side has been studying the best methods and algorithms used, and the conclusion is that the fingerprint is the best, but it has some flaws. Fingerprint algorithm has been improved so that their performance can be adapted to enhance the clarity of the edge of the gully structures of pictures fingerprint, taking into account the evaluation of the direction of the nearby edges and repeat. In the side of the computer features, computer and its components like human have unique characteristics. A program has been produced in the Visual Basic environment. The goal of this program is to get the computer characteristics and merge them with human characteristics to produce powerful algorithms of authentication and authorization can be used to protect the resources that are stored in the computer networks environments through the creation of software modules and interactive interfaces to accomplish this purpose.
Detecting and subtracting the Motion objects from backgrounds is one of the most important areas. The development of cameras and their widespread use in most areas of security, surveillance, and others made face this problem. The difficulty of this area is unstable in the classification of the pixels (foreground or background). This paper proposed a suggested background subtraction algorithm based on the histogram. The classification threshold is adaptively calculated according to many tests. The performance of the proposed algorithms was compared with state-of-the-art methods in complex dynamic scenes.
The digital image with the wavelet tools is increasing nowadays with MATLAB library, by using this method based on invariant moments which are a set of seven moments can be derived from the second and third moments , which can be calculated after converting the image from colored map to gray scale , rescale the image to (512 * 512 ) pixel , dividing the image in to four equal pieces (256 * 256 ) for each piece , then for gray scale image ( 512 * 512 ) and the four pieces (256 * 256 ) calculate wavelet with moment and invariant moment, then store the result with the author ,owner for this image to build data base for the original image to decide the authority of these images by u
... Show MoreThis paper introduces a non-conventional approach with multi-dimensional random sampling to solve a cocaine abuse model with statistical probability. The mean Latin hypercube finite difference (MLHFD) method is proposed for the first time via hybrid integration of the classical numerical finite difference (FD) formula with Latin hypercube sampling (LHS) technique to create a random distribution for the model parameters which are dependent on time [Formula: see text]. The LHS technique gives advantage to MLHFD method to produce fast variation of the parameters’ values via number of multidimensional simulations (100, 1000 and 5000). The generated Latin hypercube sample which is random or non-deterministic in nature is further integ
... Show MoreThe vast advantages of 3D modelling industry have urged competitors to improve capturing techniques and processing pipelines towards minimizing labour requirements, saving time and reducing project risk. When it comes to digital 3D documentary and conserving projects, laser scanning and photogrammetry are compared to choose between the two. Since both techniques have pros and cons, this paper approaches the potential issues of individual techniques in terms of time, budget, accuracy, density, methodology and ease to use. Terrestrial laser scanner and close-range photogrammetry are tested to document a unique invaluable artefact (Lady of Hatra) located in Iraq for future data fusion sc
To maintain the security and integrity of data, with the growth of the Internet and the increasing prevalence of transmission channels, it is necessary to strengthen security and develop several algorithms. The substitution scheme is the Playfair cipher. The traditional Playfair scheme uses a small 5*5 matrix containing only uppercase letters, making it vulnerable to hackers and cryptanalysis. In this study, a new encryption and decryption approach is proposed to enhance the resistance of the Playfair cipher. For this purpose, the development of symmetric cryptography based on shared secrets is desired. The proposed Playfair method uses a 5*5 keyword matrix for English and a 6*6 keyword matrix for Arabic to encrypt the alphabets of
... Show MoreIn this study, a fast block matching search algorithm based on blocks' descriptors and multilevel blocks filtering is introduced. The used descriptors are the mean and a set of centralized low order moments. Hierarchal filtering and MAE similarity measure were adopted to nominate the best similar blocks lay within the pool of neighbor blocks. As next step to blocks nomination the similarity of the mean and moments is used to classify the nominated blocks and put them in one of three sub-pools, each one represents certain nomination priority level (i.e., most, less & least level). The main reason of the introducing nomination and classification steps is a significant reduction in the number of matching instances of the pixels belong to the c
... Show MoreThe aim of this research is to compare traditional and modern methods to obtain the optimal solution using dynamic programming and intelligent algorithms to solve the problems of project management.
It shows the possible ways in which these problems can be addressed, drawing on a schedule of interrelated and sequential activities And clarifies the relationships between the activities to determine the beginning and end of each activity and determine the duration and cost of the total project and estimate the times used by each activity and determine the objectives sought by the project through planning, implementation and monitoring to maintain the budget assessed
... Show MoreThis study confirms the ubiquitin conjugating enzyme 2B (Rad6) plays a significant role in the DNA repair pathway also because the ubiquitin-conjugating pathway. The DNA repair pathway could be a variety of bypass repair mechanism where the broken base pair is bypassed by permitting the replication fork to labor under the site of injury. This is often done by a shift mechanism wherever deoxyribonucleic acid enzyme - δ is switched with DNA enzyme - η (DNAP - η). Site of DNAP - η is massive enough to permit the broken ester to labor under, and so bypass the broken nucleotide. However, this is often potential solely through the involvement of Proliferating cell nuclear antigen (PCNA) that could be a processivity issue and it acts as a plat
... Show More