Because of the quick growth of electrical instruments used in noxious gas detection, the importance of gas sensors has increased. X-ray diffraction (XRD) can be used to examine the crystal phase structure of sensing materials, which affects the properties of gas sensing. This contributes to the study of the effect of electrochemical synthesis of titanium dioxide (TiO2) materials with various crystal phase shapes, such as rutile TiO2 (R-TiO2NTs) and anatase TiO2 (A-TiO2NTs). In this work, we have studied the effect of voltage on preparing TiO2 nanotube arrays via the anodization technique for gas sensor applications. The results acquired from XRD, energy dispersion spectroscopy (EDX), and field emission scanning electron microscopy (FE-SEM) elucidate that TiO2 was created. In addition, systematically examining the gas detection properties was also done. The gas sensor was produced from TiO2 nanotubes, and the gas-detecting features were directed at nitrogen dioxide (NO2), which is a hazardous gas. The sensor formed from TiO2 nanotubes detects NO2 gas at various temperatures, from room temperature to 300 oC, and it has good sensitivity to this gas. The results exhibit that the gas sensor that was synthesized at 30 V has good sensitivity and a short response time at room temperature for NO2 gas sensing.
This work deals with preparation of zeolite 5A from Dewekhala kaolin clay in Al-Anbar region for drying and desulphurization of liquefied petroleum gas. The preparation of zeolite 5A includes treating kaolin clay with dilute hydrochloric acid 1N, treating metakaolin with NaOH solution to prepare 4A zeolite, ion exchange, and formation. For preparation of zeolite 4A, metakaolin treated at different temperatures (40, 60, 80, 90, and 100 °C) with different concentrations of sodium hydroxide solution (1, 2, 3, and 4 N) for 2 hours. The zeolite samples give the best relative crystallinity of zeolite prepared at 80 °C with NaOH concentration 3N (199%), and at 90 and 100°C with NaOH concentration solution 2N (184% and 189%, respectively). Ze
... Show MoreA novel median filter based on crow optimization algorithms (OMF) is suggested to reduce the random salt and pepper noise and improve the quality of the RGB-colored and gray images. The fundamental idea of the approach is that first, the crow optimization algorithm detects noise pixels, and that replacing them with an optimum median value depending on a criterion of maximization fitness function. Finally, the standard measure peak signal-to-noise ratio (PSNR), Structural Similarity, absolute square error and mean square error have been used to test the performance of suggested filters (original and improved median filter) used to removed noise from images. It achieves the simulation based on MATLAB R2019b and the resul
... Show MoreAuthentication is the process of determining whether someone or something is, in fact, who or what it is declared to be. As the dependence upon computers and computer networks grows, the need for user authentication has increased. User’s claimed identity can be verified by one of several methods. One of the most popular of these methods is represented by (something user know), such as password or Personal Identification Number (PIN). Biometrics is the science and technology of authentication by identifying the living individual’s physiological or behavioral attributes. Keystroke authentication is a new behavioral access control system to identify legitimate users via their typing behavior. The objective of this paper is to provide user
... Show MoreIris research is focused on developing techniques for identifying and locating relevant biometric features, accurate segmentation and efficient computation while lending themselves to compression methods. Most iris segmentation methods are based on complex modelling of traits and characteristics which, in turn, reduce the effectiveness of the system being used as a real time system. This paper introduces a novel parameterized technique for iris segmentation. The method is based on a number of steps starting from converting grayscale eye image to a bit plane representation, selection of the most significant bit planes followed by a parameterization of the iris location resulting in an accurate segmentation of the iris from the origin
... Show MoreThe fetal heart rate (FHR) signal processing based on Artificial Neural Networks (ANN),Fuzzy Logic (FL) and frequency domain Discrete Wavelet Transform(DWT) were analysis in order to perform automatic analysis using personal computers. Cardiotocography (CTG) is a primary biophysical method of fetal monitoring. The assessment of the printed CTG traces was based on the visual analysis of patterns that describing the variability of fetal heart rate signal. Fetal heart rate data of pregnant women with pregnancy between 38 and 40 weeks of gestation were studied. The first stage in the system was to convert the cardiotocograghy (CTG) tracing in to digital series so that the system can be analyzed ,while the second stage ,the FHR time series was t
... Show MoreArabic text categorization for pattern recognitions is challenging. We propose for the first time a novel holistic method based on clustering for classifying Arabic writer. The categorization is accomplished stage-wise. Firstly, these document images are sectioned into lines, words, and characters. Secondly, their structural and statistical features are obtained from sectioned portions. Thirdly, F-Measure is used to evaluate the performance of the extracted features and their combination in different linkage methods for each distance measures and different numbers of groups. Finally, experiments are conducted on the standard KHATT dataset of Arabic handwritten text comprised of varying samples from 1000 writers. The results in the generatio
... Show MoreToday with increase using social media, a lot of researchers have interested in topic extraction from Twitter. Twitter is an unstructured short text and messy that it is critical to find topics from tweets. While topic modeling algorithms such as Latent Semantic Analysis (LSA) and Latent Dirichlet Allocation (LDA) are originally designed to derive topics from large documents such as articles, and books. They are often less efficient when applied to short text content like Twitter. Luckily, Twitter has many features that represent the interaction between users. Tweets have rich user-generated hashtags as keywords. In this paper, we exploit the hashtags feature to improve topics learned