Iris research is focused on developing techniques for identifying and locating relevant biometric features, accurate segmentation and efficient computation while lending themselves to compression methods. Most iris segmentation methods are based on complex modelling of traits and characteristics which, in turn, reduce the effectiveness of the system being used as a real time system. This paper introduces a novel parameterized technique for iris segmentation. The method is based on a number of steps starting from converting grayscale eye image to a bit plane representation, selection of the most significant bit planes followed by a parameterization of the iris location resulting in an accurate segmentation of the iris from the original image. A lossless Hexadata encoding method is then applied to the data, which is based on reducing each set of six data items to a single encoded value. The tested results achieved acceptable saving bytes performance for the 21 iris square images of sizes 256x256 pixels which is about 22.4 KB on average with 0.79 sec decompression average time, with high saving bytes performance for 2 iris non-square images of sizes 640x480/2048x1536 that reached 76KB/2.2 sec, 1630 KB/4.71 sec respectively, Finally, the proposed promising techniques standard lossless JPEG2000 compression techniques with reduction about 1.2 and more in KB saving that implicitly demonstrating the power and efficiency of the suggested lossless biometric techniques.
The lossy-FDNR based aclive fil ter has an important property among many design realizations. 'This includes a significant reduction in component count particularly in the number of OP-AMP which consumes power. However the· problem of this type is the large component spreads which affect the fdter performance.
In this paper Genetic Algorithm is applied to minimize the component spread (capacitance and resistance p,read). The minimization of these spreads allow the fil
... Show MoreDust is a frequent contributor to health risks and changes in the climate, one of the most dangerous issues facing people today. Desertification, drought, agricultural practices, and sand and dust storms from neighboring regions bring on this issue. Deep learning (DL) long short-term memory (LSTM) based regression was a proposed solution to increase the forecasting accuracy of dust and monitoring. The proposed system has two parts to detect and monitor the dust; at the first step, the LSTM and dense layers are used to build a system using to detect the dust, while at the second step, the proposed Wireless Sensor Networks (WSN) and Internet of Things (IoT) model is used as a forecasting and monitoring model. The experiment DL system
... Show MoreVol. 6, Issue 1 (2025)
The detailed data of the Vp/Vs ratio and porosity logs were used to detect the Oil-Water Contact Zone (OWCZ) of Nahr Umr sandstone and Mishrif limestone reservoir formations in Kumiat (Kt) and Dujaila (Du) oil fields, southeastern Iraq. The results of OWC were confirmed using P-wave, Resistivity, and Water Saturation (Sw) logs of Kt-1 and Du-1 wells. It was found that the values of the oil-water contact zone thickness in Nahr Umr sandstone and Mishrif limestone were approximately one meter and eight meters, respectively. These results suggest that the OWCZ is possibly thicker in the carbonate rock than clastic rock formations. The thickness of OWCZ in the clastic rocks changed from one part to another, depending on several factors includ
... Show MorePolymer concrete were prepared by mixing epoxy resin with sand particles in three different grain size (150-300) , (300-600 ) and (600- 1200) μm respectively. The percentage of epoxy was 15%, 20 %, 25% and 30% wt of the total weight. Compression strength and flexural strength tests were carried out for the prepared samples.
The percentages of epoxy resin at 20% wt and 25% wt showed best mechanical properties for all grain sizes. These percentages were adopted to fill the voids between particles sand have two different size ranges (150-600) μm and {(150-300) & (600-1200)} μm respectively to obtain more dense material. The results showed that the strength of polymer composite at 20% resin is higher than 25% resin.
Storing and transferring the images data are raised in recent years due to requisiteness of transmission bandwidth for considerable storage capacity. Data compression method is proposed and applied in an attempt to convert data files into smaller files. The proposed and applied method is based on the Wavelet Difference Reduction (WDR) as considered the most efficient image coding method in recent years. Compression are done for three different Wavelet based Image techniques using WDR process. These techniques are implemented with different types of wavelet codecs. These are Daub2+2,2 Integer Wavelet transform, Daub5/3 integer to integer wavelet transform, and Daub9/7 Wavelet transform with level four. The used mu
... Show MoreFG Mohammed, HM Al-Dabbas, Iraqi journal of science, 2018 - Cited by 6
The objective of the current research is to find an optimum design of hybrid laminated moderate thick composite plates with static constraint. The stacking sequence and ply angle is required for optimization to achieve minimum deflection for hybrid laminated composite plates consist of glass and carbon long fibers reinforcements that impeded in epoxy matrix with known plates dimension and loading. The analysis of plate is by adopting the first-order shear deformation theory and using Navier's solution with Genetic Algorithm to approach the current objective. A program written with MATLAB to find best stacking sequence and ply angles that give minimum deflection, and the results comparing with ANSYS.
The current research aims to identify pictorial coding and its relationship to the aesthetic taste of art education students. The research community consisted of (10) plastic artworks, and (3) artworks were selected as a sample for analysis and decoding. With the aim of the research, the research tool was prepared as it consisted of an analysis form, and the researcher used statistical methods: Equation (Cooper) to find the percentage of agreement between the arbitrators and the equation (Scott) to calculate the validity of the tool, and the correlation coefficient (Pearson) to extract stability in the method of segmentation half. Shape formations and achieve encryption of the plastic image through decoding symbols, meanings, and the sig
... Show MoreThe present paper concerns with the problem of estimating the reliability system in the stress – strength model under the consideration non identical and independent of stress and strength and follows Lomax Distribution. Various shrinkage estimation methods were employed in this context depend on Maximum likelihood, Moment Method and shrinkage weight factors based on Monte Carlo Simulation. Comparisons among the suggested estimation methods have been made using the mean absolute percentage error criteria depend on MATLAB program.