In this paper, we deal with the problem of general matching of two images one of them has experienced geometrical transformations, to find the correspondence between two images. We develop the invariant moments for traditional techniques (moments of inertia) with new approach to enhance the performance for these methods. We test various projections directional moments, to extract the difference between Block Distance Moment (BDM) and evaluate their reliability. Three adaptive strategies are shown for projections directional moments, that are raster (vertical and horizontal) projection, Fan-Bean projection and new projection procedure that is the square projection method. Our paper started with the description of a new algorithm that is low
... Show More
The research aims to study and analysis of concurrent engineering (CE) and cost optimization (CO), and the use of concurrent engineering inputs to outputs to improve the cost, and the statement of the role of concurrent engineering in improving the quality of the product, and achieve savings in the design and manufacturing time and assembly and reduce costs, as well as employing some models to determine how much the savings in time, including the model (Lexmark) model (Pert) to determine the savings in design time for manufacturing and assembly time.
To achieve the search objectives, the General Company for Electrical and Electronic Industries \ Refrigerated Engine
... Show MoreTwo simple and sensitive spectrophotometric methods are proposed for the determination of amitriptyline in its pure form and in tablets. The first method is based on the formation of charge- transfer complex between amitriptyline as n-donor and tetracyano-ethylene (TCNE) as πacceptor. The product exhibit absorbance maximum at 470 nm in acetonitrile solvent (pH =9.0 ) . In the second method the absorbance of the ion- pair complex, which is formed between the soughted drug and bromocresol green (BCG), was measured at 415 nm at ( pH=3.5) . In addition to classical univariate optimization, modified simplex method (MSM) was applied in the optimization of the variable affecting the color producing reaction by a geometric simple
... Show MoreAbstract
The logistic regression model is one of the nonlinear models that aims at obtaining highly efficient capabilities, It also the researcher an idea of the effect of the explanatory variable on the binary response variable. &nb
... Show MoreObjective of this work is the mixing between human biometric characteristics and unique attributes of the computer in order to protect computer networks and resources environments through the development of authentication and authorization techniques. In human biometric side has been studying the best methods and algorithms used, and the conclusion is that the fingerprint is the best, but it has some flaws. Fingerprint algorithm has been improved so that their performance can be adapted to enhance the clarity of the edge of the gully structures of pictures fingerprint, taking into account the evaluation of the direction of the nearby edges and repeat. In the side of the computer features, computer and its components like human have uniqu
... Show MoreThis paper include the problem of segmenting an image into regions represent (objects), segment this object by define boundary between two regions using a connected component labeling. Then develop an efficient segmentation algorithm based on this method, to apply the algorithm to image segmentation using different kinds of images, this algorithm consist four steps at the first step convert the image gray level the are applied on the image, these images then in the second step convert to binary image, edge detection using Canny edge detection in third Are applie the final step is images. Best segmentation rates are (90%) obtained when using the developed algorithm compared with (77%) which are obtained using (ccl) before enhancement.
In this paper, the effective computational method (ECM) based on the standard monomial polynomial has been implemented to solve the nonlinear Jeffery-Hamel flow problem. Moreover, novel effective computational methods have been developed and suggested in this study by suitable base functions, namely Chebyshev, Bernstein, Legendre, and Hermite polynomials. The utilization of the base functions converts the nonlinear problem to a nonlinear algebraic system of equations, which is then resolved using the Mathematica®12 program. The development of effective computational methods (D-ECM) has been applied to solve the nonlinear Jeffery-Hamel flow problem, then a comparison between the methods has been shown. Furthermore, the maximum
... Show MoreCryptography is the process of transforming message to avoid an unauthorized access of data. One of the main problems and an important part in cryptography with secret key algorithms is key. For higher level of secure communication key plays an important role. For increasing the level of security in any communication, both parties must have a copy of the secret key which, unfortunately, is not that easy to achieve. Triple Data Encryption Standard algorithm is weak due to its weak key generation, so that key must be reconfigured to make this algorithm more secure, effective, and strong. Encryption key enhances the Triple Data Encryption Standard algorithm securities. This paper proposed a combination of two efficient encryption algorithms to
... Show MoreIn this paper, a new modification was proposed to enhance the security level in the Blowfish algorithm by increasing the difficulty of cracking the original message which will lead to be safe against unauthorized attack. This algorithm is a symmetric variable-length key, 64-bit block cipher and it is implemented using gray scale images of different sizes. Instead of using a single key in cipher operation, another key (KEY2) of one byte length was used in the proposed algorithm which has taken place in the Feistel function in the first round both in encryption and decryption processes. In addition, the proposed modified Blowfish algorithm uses five Sboxes instead of four; the additional key (KEY2) is selected randomly from additional Sbox
... Show MoreVariable selection is an essential and necessary task in the statistical modeling field. Several studies have triedto develop and standardize the process of variable selection, but it isdifficultto do so. The first question a researcher needs to ask himself/herself what are the most significant variables that should be used to describe a given dataset’s response. In thispaper, a new method for variable selection using Gibbs sampler techniqueshas beendeveloped.First, the model is defined, and the posterior distributions for all the parameters are derived.The new variable selection methodis tested usingfour simulation datasets. The new approachiscompared with some existingtechniques: Ordinary Least Squared (OLS), Least Absolute Shrinkage
... Show More