Document clustering is the process of organizing a particular electronic corpus of documents into subgroups of similar text features. Formerly, a number of conventional algorithms had been applied to perform document clustering. There are current endeavors to enhance clustering performance by employing evolutionary algorithms. Thus, such endeavors became an emerging topic gaining more attention in recent years. The aim of this paper is to present an up-to-date and self-contained review fully devoted to document clustering via evolutionary algorithms. It firstly provides a comprehensive inspection to the document clustering model revealing its various components with its related concepts. Then it shows and analyzes the principle research work in this topic. Finally, it compiles and classifies various objective functions, the core of the evolutionary algorithms, from the related collection of research papers. The paper ends up by addressing some important issues and challenges that can be subject of future work.
The rapid and enormous growth of the Internet of Things, as well as its widespread adoption, has resulted in the production of massive quantities of data that must be processed and sent to the cloud, but the delay in processing the data and the time it takes to send it to the cloud has resulted in the emergence of fog, a new generation of cloud in which the fog serves as an extension of cloud services at the edge of the network, reducing latency and traffic. The distribution of computational resources to minimize makespan and running costs is one of the disadvantages of fog computing. This paper provides a new approach for improving the task scheduling problem in a Cloud-Fog environme
Breast cancer is the second deadliest disease infected women worldwide. For this
reason the early detection is one of the most essential stop to overcomeit dependingon
automatic devices like artificial intelligent. Medical applications of machine learning
algorithmsare mostly based on their ability to handle classification problems,
including classifications of illnesses or to estimate prognosis. Before machine
learningis applied for diagnosis, it must be trained first. The research methodology
which isdetermines differentofmachine learning algorithms,such as Random tree,
ID3, CART, SMO, C4.5 and Naive Bayesto finds the best training algorithm result.
The contribution of this research is test the data set with mis
The expanding use of multi-processor supercomputers has made a significant impact on the speed and size of many problems. The adaptation of standard Message Passing Interface protocol (MPI) has enabled programmers to write portable and efficient codes across a wide variety of parallel architectures. Sorting is one of the most common operations performed by a computer. Because sorted data are easier to manipulate than randomly ordered data, many algorithms require sorted data. Sorting is of additional importance to parallel computing because of its close relation to the task of routing data among processes, which is an essential part of many parallel algorithms. In this paper, sequential sorting algorithms, the parallel implementation of man
... Show MoreBrain tissues segmentation is usually concerned with the delineation of three types of brain matters Grey Matter (GM), White Matter (WM) and Cerebrospinal Fluid (CSF). Because most brain structures are anatomically defined by boundaries of these tissue classes, accurate segmentation of brain tissues into one of these categories is an important step in quantitative morphological study of the brain. As well as the abnormalities regions like tumors are needed to be delineated. The extra-cortical voxels in MR brain images are often removed in order to facilitate accurate analysis of cortical structures. Brain extraction is necessary to avoid the misclassifications of surrounding tissues, skull and scalp as WM, GM or tumor when implementing s
... Show MoreOne of the significant stages in computer vision is image segmentation which is fundamental for different applications, for example, robot control and military target recognition, as well as image analysis of remote sensing applications. Studies have dealt with the process of improving the classification of all types of data, whether text or audio or images, one of the latest studies in which researchers have worked to build a simple, effective, and high-accuracy model capable of classifying emotions from speech data, while several studies dealt with improving textual grouping. In this study, we seek to improve the classification of image division using a novel approach depending on two methods used to segment the images. The first
... Show MoreIn the recent years, remote sensing applications have a great interest because it's offers many advantages, benefits and possibilities for the applications that using this concept, satellite it's one must important applications for remote sensing, it's provide us with multispectral images allow as study many problems like changing in ecological cover or biodiversity for earth surfers, and illustrated biological diversity of the studied areas by the presentation of the different areas of the scene taken depending on the length of the characteristic wave, Thresholding it's a common used operation for image segmentation, it's seek to extract a monochrome image from gray image by segment this image to two region (for
... Show MoreIn this paper, new method have been investigated using evolving algorithms (EA's) to cryptanalysis one of the nonlinear stream cipher cryptosystems which depends on the Linear Feedback Shift Register (LFSR) unit by using cipher text-only attack. Genetic Algorithm (GA) and Ant Colony Optimization (ACO) which are used for attacking one of the nonlinear cryptosystems called "shrinking generator" using different lengths of cipher text and different lengths of combined LFSRs. GA and ACO proved their good performance in finding the initial values of the combined LFSRs. This work can be considered as a warning for a stream cipher designer to avoid the weak points, which may be f
... Show MoreAlthough the number of stomach tumor patients reduced obviously during last decades in western countries, but this illness is still one of the main causes of death in developing countries. The aim of this research is to detect the area of a tumor in a stomach images based on fuzzy clustering. The proposed methodology consists of three stages. The stomach images are divided into four quarters and then features elicited from each quarter in the first stage by utilizing seven moments invariant. Fuzzy C-Mean clustering (FCM) was employed in the second stage for each quarter to collect the features of each quarter into clusters. Manhattan distance was calculated in the third stage among all clusters' centers in all quarters to disclosure of t
... Show More