<span>One of the main difficulties facing the certified documents documentary archiving system is checking the stamps system, but, that stamps may be contains complex background and surrounded by unwanted data. Therefore, the main objective of this paper is to isolate background and to remove noise that may be surrounded stamp. Our proposed method comprises of four phases, firstly, we apply k-means algorithm for clustering stamp image into a number of clusters and merged them using ISODATA algorithm. Secondly, we compute mean and standard deviation for each remaining cluster to isolate background cluster from stamp cluster. Thirdly, a region growing algorithm is applied to segment the image and then choosing the connected region to produce a binary mask for the stamp area. Finally, the binary mask is combined with the original image to extract the stamp regions. The results indicate that the number of clusters can be determined dynamically and the largest cluster that has minimum standard deviation (i.e., always the largest cluster is the background cluster). Also, show that the binary mask can be established from more than one segment to cover are all stamp’s disconnected pieces and it can be useful to remove the noise appear with stamp region.</span>
In this paper a nonlinear adaptive control method is presented for a pH process, which is difficult to control due to the nonlinear and uncertainties. A theoretical and experimental investigation was conducted of the dynamic behavior of neutralization process in a continuous stirred tank reactor (CSTR). The process control was implemented using different control strategies, velocity form of PI control and nonlinear adaptive control. Through simulation studies it has been shown that the estimated parameters are in good agreement with the actual values and that the proposed adaptive controller has excellent tracking and regulation performance.
In this paper, a hybrid image compression technique is introduced that integrates discrete wavelet transform (DWT) and linear polynomial coding. In addition, the proposed technique improved the midtread quantizer scheme once by utilizing the block based and the selected factor value. The compression system performance showed the superiority in quality and compression ratio compared to traditional polynomial coding techniques.
Data generated from modern applications and the internet in healthcare is extensive and rapidly expanding. Therefore, one of the significant success factors for any application is understanding and extracting meaningful information using digital analytics tools. These tools will positively impact the application's performance and handle the challenges that can be faced to create highly consistent, logical, and information-rich summaries. This paper contains three main objectives: First, it provides several analytics methodologies that help to analyze datasets and extract useful information from them as preprocessing steps in any classification model to determine the dataset characteristics. Also, this paper provides a comparative st
... Show MoreDocument clustering is the process of organizing a particular electronic corpus of documents into subgroups of similar text features. Formerly, a number of conventional algorithms had been applied to perform document clustering. There are current endeavors to enhance clustering performance by employing evolutionary algorithms. Thus, such endeavors became an emerging topic gaining more attention in recent years. The aim of this paper is to present an up-to-date and self-contained review fully devoted to document clustering via evolutionary algorithms. It firstly provides a comprehensive inspection to the document clustering model revealing its various components with its related concepts. Then it shows and analyzes the principle research wor
... Show MoreIn this article, results have been shown via using a general quasi contraction multi-valued mapping in Cat(0) space. These results are used to prove the convergence of two iteration algorithms to a fixed point and the equivalence of convergence. We also demonstrate an appropriate conditions to ensure that one is faster than others.
The purpose of this paper is to solve the stochastic demand for the unbalanced transport problem using heuristic algorithms to obtain the optimum solution, by minimizing the costs of transporting the gasoline product for the Oil Products Distribution Company of the Iraqi Ministry of Oil. The most important conclusions that were reached are the results prove the possibility of solving the random transportation problem when the demand is uncertain by the stochastic programming model. The most obvious finding to emerge from this work is that the genetic algorithm was able to address the problems of unbalanced transport, And the possibility of applying the model approved by the oil products distribution company in the Iraqi Ministry of Oil to m
... Show MoreThe aim of this paper, is to discuss several high performance training algorithms fall into two main categories. The first category uses heuristic techniques, which were developed from an analysis of the performance of the standard gradient descent algorithm. The second category of fast algorithms uses standard numerical optimization techniques such as: quasi-Newton . Other aim is to solve the drawbacks related with these training algorithms and propose an efficient training algorithm for FFNN
Feature selection represents one of the critical processes in machine learning (ML). The fundamental aim of the problem of feature selection is to maintain performance accuracy while reducing the dimension of feature selection. Different approaches were created for classifying the datasets. In a range of optimization problems, swarming techniques produced better outcomes. At the same time, hybrid algorithms have gotten a lot of attention recently when it comes to solving optimization problems. As a result, this study provides a thorough assessment of the literature on feature selection problems using hybrid swarm algorithms that have been developed over time (2018-2021). Lastly, when compared with current feature selection procedu
... Show MoreToday with increase using social media, a lot of researchers have interested in topic extraction from Twitter. Twitter is an unstructured short text and messy that it is critical to find topics from tweets. While topic modeling algorithms such as Latent Semantic Analysis (LSA) and Latent Dirichlet Allocation (LDA) are originally designed to derive topics from large documents such as articles, and books. They are often less efficient when applied to short text content like Twitter. Luckily, Twitter has many features that represent the interaction between users. Tweets have rich user-generated hashtags as keywords. In this paper, we exploit the hashtags feature to improve topics learned
Eye loss may be caused as a result of eye trauma, accidents, or malignant tumors, which leads the patient to undergo surgery to remove the damaged parts. This research examines the potential of computer vision represented by Structure from Motion (SfM) photogrammetry in fabricating the orbital prosthesis as a noninvasive and low-cost technique. A low-cost camera was used to collect the data towards extracting the dense 3D data of the patient facial features following Structure from Motion-Multi View Stereo (SfM-MVS) algorithms. To restore the defective orbital, a Reverse Engineering (RE) based approach has been applied using the similarity RE algorithms based on the opposite healthy eye to rehabilitate the defected orbital precisely
... Show More