Document clustering is the process of organizing a particular electronic corpus of documents into subgroups of similar text features. Formerly, a number of conventional algorithms had been applied to perform document clustering. There are current endeavors to enhance clustering performance by employing evolutionary algorithms. Thus, such endeavors became an emerging topic gaining more attention in recent years. The aim of this paper is to present an up-to-date and self-contained review fully devoted to document clustering via evolutionary algorithms. It firstly provides a comprehensive inspection to the document clustering model revealing its various components with its related concepts. Then it shows and analyzes the principle research work in this topic. Finally, it compiles and classifies various objective functions, the core of the evolutionary algorithms, from the related collection of research papers. The paper ends up by addressing some important issues and challenges that can be subject of future work.
Evolutionary algorithms are better than heuristic algorithms at finding protein complexes in protein-protein interaction networks (PPINs). Many of these algorithms depend on their standard frameworks, which are based on topology. Further, many of these algorithms have been exclusively examined on networks with only reliable interaction data. The main objective of this paper is to extend the design of the canonical and topological-based evolutionary algorithms suggested in the literature to cope with noisy PPINs. The design of the evolutionary algorithm is extended based on the functional domain of the proteins rather than on the topological domain of the PPIN. The gene ontology annotation in each molecular function, biological proce
... Show MoreThe rapid and enormous growth of the Internet of Things, as well as its widespread adoption, has resulted in the production of massive quantities of data that must be processed and sent to the cloud, but the delay in processing the data and the time it takes to send it to the cloud has resulted in the emergence of fog, a new generation of cloud in which the fog serves as an extension of cloud services at the edge of the network, reducing latency and traffic. The distribution of computational resources to minimize makespan and running costs is one of the disadvantages of fog computing. This paper provides a new approach for improving the task scheduling problem in a Cloud-Fog environme
Software-defined networks (SDN) have a centralized control architecture that makes them a tempting target for cyber attackers. One of the major threats is distributed denial of service (DDoS) attacks. It aims to exhaust network resources to make its services unavailable to legitimate users. DDoS attack detection based on machine learning algorithms is considered one of the most used techniques in SDN security. In this paper, four machine learning techniques (Random Forest, K-nearest neighbors, Naive Bayes, and Logistic Regression) have been tested to detect DDoS attacks. Also, a mitigation technique has been used to eliminate the attack effect on SDN. RF and KNN were selected because of their high accuracy results. Three types of ne
... Show MoreOne of the significant stages in computer vision is image segmentation which is fundamental for different applications, for example, robot control and military target recognition, as well as image analysis of remote sensing applications. Studies have dealt with the process of improving the classification of all types of data, whether text or audio or images, one of the latest studies in which researchers have worked to build a simple, effective, and high-accuracy model capable of classifying emotions from speech data, while several studies dealt with improving textual grouping. In this study, we seek to improve the classification of image division using a novel approach depending on two methods used to segment the images. The first
... Show MoreThe segmentation of aerial images using different clustering techniques offers valuable insights into interpreting and analyzing such images. By partitioning the images into meaningful regions, clustering techniques help identify and differentiate various objects and areas of interest, facilitating various applications, including urban planning, environmental monitoring, and disaster management. This paper aims to segment color aerial images to provide a means of organizing and understanding the visual information contained within the image for various applications and research purposes. It is also important to look into and compare the basic workings of three popular clustering algorithms: K-Medoids, Fuzzy C-Mean (FCM), and Gaussia
... Show MoreCurrently, there is an intensive development of bipedal walking robots. The most known solutions are based on the use of the principles of human gait created in nature during evolution. Modernbipedal robots are also based on the locomotion manners of birds. This review presents the current state of the art of bipedal walking robots based on natural bipedal movements (human and bird) as well as on innovative synthetic solutions. Firstly, an overview of the scientific analysis of human gait is provided as a basis for the design of bipedal robots. The full human gait cycle that consists of two main phases is analysed and the attention is paid to the problem of balance and stability, especially in the single support phase when the biped
... Show MoreAlthough the number of stomach tumor patients reduced obviously during last decades in western countries, but this illness is still one of the main causes of death in developing countries. The aim of this research is to detect the area of a tumor in a stomach images based on fuzzy clustering. The proposed methodology consists of three stages. The stomach images are divided into four quarters and then features elicited from each quarter in the first stage by utilizing seven moments invariant. Fuzzy C-Mean clustering (FCM) was employed in the second stage for each quarter to collect the features of each quarter into clusters. Manhattan distance was calculated in the third stage among all clusters' centers in all quarters to disclosure of t
... Show MoreThe expanding use of multi-processor supercomputers has made a significant impact on the speed and size of many problems. The adaptation of standard Message Passing Interface protocol (MPI) has enabled programmers to write portable and efficient codes across a wide variety of parallel architectures. Sorting is one of the most common operations performed by a computer. Because sorted data are easier to manipulate than randomly ordered data, many algorithms require sorted data. Sorting is of additional importance to parallel computing because of its close relation to the task of routing data among processes, which is an essential part of many parallel algorithms. In this paper, sequential sorting algorithms, the parallel implementation of man
... Show More