Numeral recognition is considered an essential preliminary step for optical character recognition, document understanding, and others. Although several handwritten numeral recognition algorithms have been proposed so far, achieving adequate recognition accuracy and execution time remain challenging to date. In particular, recognition accuracy depends on the features extraction mechanism. As such, a fast and robust numeral recognition method is essential, which meets the desired accuracy by extracting the features efficiently while maintaining fast implementation time. Furthermore, to date most of the existing studies are focused on evaluating their methods based on clean environments, thus limiting understanding of their potential application in more realistic noise environments. Therefore, finding a feasible and accurate handwritten numeral recognition method that is accurate in the more practical noisy environment is crucial. To this end, this paper proposes a new scheme for handwritten numeral recognition using Hybrid orthogonal polynomials. Gradient and smoothed features are extracted using the hybrid orthogonal polynomial. To reduce the complexity of feature extraction, the embedded image kernel technique has been adopted. In addition, support vector machine is used to classify the extracted features for the different numerals. The proposed scheme is evaluated under three different numeral recognition datasets: Roman, Arabic, and Devanagari. We compare the accuracy of the proposed numeral recognition method with the accuracy achieved by the state-of-the-art recognition methods. In addition, we compare the proposed method with the most updated method of a convolutional neural network. The results show that the proposed method achieves almost the highest recognition accuracy in comparison with the existing recognition methods in all the scenarios considered. Importantly, the results demonstrate that the proposed method is robust against the noise distortion and outperforms the convolutional neural network considerably, which signifies the feasibility and the effectiveness of the proposed approach in comparison to the state-of-the-art recognition methods under both clean noise and more realistic noise environments.
De-waxing of lubricating oil distillate (400-500 ºC) by using urea was investigated in the present study. Lubricating oil distillate produced by vacuum distillation and refined by furfural extraction was taken from Al-Daura refinery. This oil distillate has a pour point of 34 ºC. Two solvents were used to dilute the oil distillate, these are methyl isobutyl ketone and methylene chloride. The operating conditions of the urea adduct formation with n-paraffins in the presence of methyl isobutyl ketone were studied in details, these are solvent to oil volume ratio within the range of 0 to 2, mixer speed 0 to 2000 rpm, urea to wax weight ratio 0 to 6.3, time of adduction 0 to 71 min and temperature 30-70 ºC). Pour point of de-waxed oil and yi
... Show MorePalm vein recognition is a one of the most efficient biometric technologies, each individual can be identified through its veins unique characteristics, palm vein acquisition techniques is either contact based or contactless based, as the individual's hand contact or not the peg of the palm imaging device, the needs a contactless palm vein system in modern applications rise tow problems, the pose variations (rotation, scaling and translation transformations) since the imaging device cannot aligned correctly with the surface of the palm, and a delay of matching process especially for large systems, trying to solve these problems. This paper proposed a pose invariant identification system for contactless palm vein which include three main
... Show MoreIn the light of the globalization Which surrounds the business environment and whose impact has been reflected on industrial economic units the whole world has become a single market that affects its variables on all units and is affected by the economic contribution of each economic unit as much as its share. The problem of this research is that the use of Pareto analysis enables industrial economic units to diagnose the risks surrounding them , so the main objective of the research was to classify risks into both internal and external types and identify any risks that require more attention.
The research was based on the hypothesis that Pareto analysis used, risks can be identified and addressed before they occur.
... Show MoreThe process of accurate localization of the basic components of human faces (i.e., eyebrows, eyes, nose, mouth, etc.) from images is an important step in face processing techniques like face tracking, facial expression recognition or face recognition. However, it is a challenging task due to the variations in scale, orientation, pose, facial expressions, partial occlusions and lighting conditions. In the current paper, a scheme includes the method of three-hierarchal stages for facial components extraction is presented; it works regardless of illumination variance. Adaptive linear contrast enhancement methods like gamma correction and contrast stretching are used to simulate the variance in light condition among images. As testing material
... Show MoreThe Assignment model is a mathematical model that aims to express a real problem facing factories and companies which is characterized by the guarantee of its activity in order to make the appropriate decision to get the best allocation of machines or jobs or workers on machines in order to increase efficiency or profits to the highest possible level or reduce costs or time To the extent possible, and in this research has been using the method of labeling to solve the problem of the fuzzy assignment of real data has been approved by the tire factory Diwaniya, where the data included two factors are the factors of efficiency and cost, and was solved manually by a number of iterations until reaching the optimization solution,
... Show MoreThis paper shows an approach for Electromyography (ECG) signal processing based on linear and nonlinear adaptive filtering using Recursive Least Square (RLS) algorithm to remove two kinds of noise that affected the ECG signal. These are the High Frequency Noise (HFN) and Low Frequency Noise (LFN). Simulation is performed in Matlab. The ECG, HFN and LFN signals used in this study were downloaded from ftp://ftp.ieee.org/uploads/press/rangayyan/, and then the filtering process was obtained by using adaptive finite impulse response (FIR) that illustrated better results than infinite impulse response (IIR) filters did.
This study investigated the treatment of dairy wastewater using the electrocoagulation method with iron filings as electrodes. The study dealt with real samples collected from local factory for dairy products in Baghdad. The Response Surface Methodology (RSM) was used to optimize five experimental variables at six levels for each variable, for estimating chemical oxygen demand (COD) removal efficiency. These variables were the distance between electrodes, detention time, dosage of NaCl as electrolyte, initial COD concentration, and current density. RSM was investigated the direct and complex interaction effects between parameters to estimate the optimum values. The respective optimum value was 1 cm for the distance between electrodes, (6
... Show More<p>Generally, The sending process of secret information via the transmission channel or any carrier medium is not secured. For this reason, the techniques of information hiding are needed. Therefore, steganography must take place before transmission. To embed a secret message at optimal positions of the cover image under spatial domain, using the developed particle swarm optimization algorithm (Dev.-PSO) to do that purpose in this paper based on Least Significant Bits (LSB) using LSB substitution. The main aim of (Dev. -PSO) algorithm is determining an optimal paths to reach a required goals in the specified search space based on disposal of them, using (Dev.-PSO) algorithm produces the paths of a required goals with most effi
... Show MoreCOVID 19 has spread rapidly around the world due to the lack of a suitable vaccine; therefore the early prediction of those infected with this virus is extremely important attempting to control it by quarantining the infected people and giving them possible medical attention to limit its spread. This work suggests a model for predicting the COVID 19 virus using feature selection techniques. The proposed model consists of three stages which include the preprocessing stage, the features selection stage, and the classification stage. This work uses a data set consists of 8571 records, with forty features for patients from different countries. Two feature selection techniques are used in