In this research, the removal of cadmium (Cd) from simulated wastewater was investigated by using a fixed bed bio-electrochemical reactor. The effects of the main controlling factors on the performance of the removal process such as applied cell voltage, initial Cd concentration, pH of the catholyte, and the mesh number of the cathode were investigated. The results showed that the applied cell voltage had the main impact on the removal efficiency of cadmium where increasing the applied voltage led to higher removal efficiency. Meanwhile increasing the applied voltage was found to be given lower current efficiency and higher energy consumption. No significant effect of initial Cd concentration on the removal efficiency of cadmium but increasing the initial concentration would be given higher current efficiency and lower energy consumption. The results established that using a pH value lower than three results in a sharp decrease in the removal efficiency as well as using a pH value higher than seven results in decreasing the removal efficiency. Using a mesh number higher than 30 gave a lower removal efficiency. The best operating conditions were found to be an applied potential of 1.8 V, an initial Cd concentration of 125 ppm, and a pH of 7. Under these operating conditions with the using a stack of stainless with mesh number 30 as a packed bed cathode, a complete removal efficiency of Cd(100%) was obtained at a current efficiency of 83.57% and energy consumption of 0.57 kWh/kg Cd.
This dissertation depends on study of the topological structure in graph theory as well as introduce some concerning concepts, and generalization them into new topological spaces constructed using elements of graph. Thus, it is required presenting some theorems, propositions, and corollaries that are available in resources and proof which are not available. Moreover, studying some relationships between many concepts and examining their equivalence property like locally connectedness, convexity, intervals, and compactness. In addition, introducing the concepts of weaker separation axioms in α-topological spaces than the standard once like, α-feebly Hausdorff, α-feebly regular, and α-feebly normal and studying their properties. Furthermor
... Show MoreDocument analysis of images snapped by camera is a growing challenge. These photos are often poor-quality compound images, composed of various objects and text; this makes automatic analysis complicated. OCR is one of the image processing techniques which is used to perform automatic identification of texts. Existing image processing techniques need to manage many parameters in order to clearly recognize the text in such pictures. Segmentation is regarded one of these essential parameters. This paper discusses the accuracy of segmentation process and its effect over the recognition process. According to the proposed method, the images were firstly filtered using the wiener filter then the active contour algorithm could b
... Show MoreIn this paper we find the exact solution of Burger's equation after reducing it to Bernoulli equation. We compare this solution with that given by Kaya where he used Adomian decomposition method, the solution given by chakrone where he used the Variation iteration method (VIM)and the solution given by Eq(5)in the paper of M. Javidi. We notice that our solution is better than their solutions.
In cognitive radio networks, there are two important probabilities; the first probability is important to primary users called probability of detection as it indicates their protection level from secondary users, and the second probability is important to the secondary users called probability of false alarm which is used for determining their using of unoccupied channel. Cooperation sensing can improve the probabilities of detection and false alarm. A new approach of determine optimal value for these probabilities, is supposed and considered to face multi secondary users through discovering an optimal threshold value for each unique detection curve then jointly find the optimal thresholds. To get the aggregated throughput over transmission
... Show MoreIndustrial effluents loaded with heavy metals are a cause of hazards to the humans and other forms of life. Conventional approaches, such as electroplating, ion exchange, and membrane processes, are used for removal of copper, cadmium, and lead and are often cost prohibitive with low efficiency at low metal ion concentration. Biosorption can be considered as an option which has been proven as more efficient and economical for removing the mentioned metal ions. Biosorbents used are fungi, yeasts, oil palm shells, coir pith carbon, peanut husks, and olive pulp. Recently, low cost and natural products have also been researched as biosorbent. This paper presents an attempt of the potential use of Iraqi date pits and Al-Khriet (i.e. substances l
... Show MoreIn this paper, new method have been investigated using evolving algorithms (EA's) to cryptanalysis one of the nonlinear stream cipher cryptosystems which depends on the Linear Feedback Shift Register (LFSR) unit by using cipher text-only attack. Genetic Algorithm (GA) and Ant Colony Optimization (ACO) which are used for attacking one of the nonlinear cryptosystems called "shrinking generator" using different lengths of cipher text and different lengths of combined LFSRs. GA and ACO proved their good performance in finding the initial values of the combined LFSRs. This work can be considered as a warning for a stream cipher designer to avoid the weak points, which may be f
... Show MoreThe area of character recognition has received a considerable attention by researchers all over the world during the last three decades. However, this research explores best sets of feature extraction techniques and studies the accuracy of well-known classifiers for Arabic numeral using the Statistical styles in two methods and making comparison study between them. First method Linear Discriminant function that is yield results with accuracy as high as 90% of original grouped cases correctly classified. In the second method, we proposed algorithm, The results show the efficiency of the proposed algorithms, where it is found to achieve recognition accuracy of 92.9% and 91.4%. This is providing efficiency more than the first method.
Smart water flooding (low salinity water flooding) was mainly invested in a sandstone reservoir. The main reasons for using low salinity water flooding are; to improve oil recovery and to give a support for the reservoir pressure.
In this study, two core plugs of sandstone were used with different permeability from south of Iraq to explain the effect of water injection with different ions concentration on the oil recovery. Water types that have been used are formation water, seawater, modified low salinity water, and deionized water.
The effects of water salinity, the flow rate of water injected, and the permeability of core plugs have been studied in order to summarize the best conditions of low salinity
... Show More