A substantial portion of today’s multimedia data exists in the form of unstructured text. However, the unstructured nature of text poses a significant task in meeting users’ information requirements. Text classification (TC) has been extensively employed in text mining to facilitate multimedia data processing. However, accurately categorizing texts becomes challenging due to the increasing presence of non-informative features within the corpus. Several reviews on TC, encompassing various feature selection (FS) approaches to eliminate non-informative features, have been previously published. However, these reviews do not adequately cover the recently explored approaches to TC problem-solving utilizing FS, such as optimization techniques. This study comprehensively analyzes different FS approaches based on optimization algorithms for TC. We begin by introducing the primary phases involved in implementing TC. Subsequently, we explore a wide range of FS approaches for categorizing text documents and attempt to organize the existing works into four fundamental approaches: filter, wrapper, hybrid, and embedded. Furthermore, we review four optimization algorithms utilized in solving text FS problems: swarm intelligence-based, evolutionary-based, physics-based, and human behavior-related algorithms. We discuss the advantages and disadvantages of state-of-the-art studies that employ optimization algorithms for text FS methods. Additionally, we consider several aspects of each proposed method and thoroughly discuss the challenges associated with datasets, FS approaches, optimization algorithms, machine learning classifiers, and evaluation criteria employed to assess new and existing techniques. Finally, by identifying research gaps and proposing future directions, our review provides valuable guidance to researchers in developing and situating further studies within the current body of literature.
This work aims to develop a secure lightweight cipher algorithm for constrained devices. A secure communication among constrained devices is a critical issue during the data transmission from the client to the server devices. Lightweight cipher algorithms are defined as a secure solution for constrained devices that require low computational functions and small memory. In contrast, most lightweight algorithms suffer from the trade-off between complexity and speed in order to produce robust cipher algorithm. The PRESENT cipher has been successfully experimented on as a lightweight cryptography algorithm, which transcends other ciphers in terms of its computational processing that required low complexity operations. The mathematical model of
... Show MoreA Multiple System Biometric System Based on ECG Data
The aim of this work is to design an algorithm which combines between steganography andcryptography that can hide a text in an image in a way that prevents, as much as possible, anysuspicion of the hidden textThe proposed system depends upon preparing the image data for the next step (DCT Quantization)through steganographic process and using two levels of security: the RSA algorithm and the digitalsignature, then storing the image in a JPEG format. In this case, the secret message will be looked asplaintext with digital signature while the cover is a coloured image. Then, the results of the algorithmare submitted to many criteria in order to be evaluated that prove the sufficiency of the algorithm andits activity. Thus, the proposed algorit
... Show MoreImplementation of TSFS (Transposition, Substitution, Folding, and Shifting) algorithm as an encryption algorithm in database security had limitations in character set and the number of keys used. The proposed cryptosystem is based on making some enhancements on the phases of TSFS encryption algorithm by computing the determinant of the keys matrices which affects the implementation of the algorithm phases. These changes showed high security to the database against different types of security attacks by achieving both goals of confusion and diffusion.
The digital communication of a product of communication and information revolution. It is characterized by accurate and comprehensive in its services and its effects, which brought changes in the structure of many communities and their organizational structures. They have significant impacts on the social systems and social relations, especially in the Arab societies, which are the focus of the globalized Western media, for many reasons: economical, political , cultural and social.
According to this perception, the Arab identity has become in an encounter with big challenges by the globalized media of trade and the media, which aims to achieve greater profits because of identity and its importance to the communities. This occurs par
Language as a means of communication has long been the concern of many conversation analysts in their studies such as: Sacks et al. (1974), Schegloff et al. (1977), Duncan (1972), Grice (1975) and Burton (1980). Burton has attempted analyzing the first ten transitions of the play “The Dumb Waiter” for mere a presentation of her approach. This paper aims at analyzing the conversational structure of forum on the subject of literary fiction and genre fiction by applying Burton’s model (1980) of analysis to answer the question to what extent this model is applicable in analyzing the presented text. The findings of the investigation have proved the applicability of the structure of conversation formulated by Burton (1980) in her model wit
... Show MoreAnomaly detection is still a difficult task. To address this problem, we propose to strengthen DBSCAN algorithm for the data by converting all data to the graph concept frame (CFG). As is well known that the work DBSCAN method used to compile the data set belong to the same species in a while it will be considered in the external behavior of the cluster as a noise or anomalies. It can detect anomalies by DBSCAN algorithm can detect abnormal points that are far from certain set threshold (extremism). However, the abnormalities are not those cases, abnormal and unusual or far from a specific group, There is a type of data that is do not happen repeatedly, but are considered abnormal for the group of known. The analysis showed DBSCAN using the
... Show MoreAlthough text document images authentication is difficult due to the binary nature and clear separation between the background and foreground but it is getting higher demand for many applications. Most previous researches in this field depend on insertion watermark in the document, the drawback in these techniques lie in the fact that changing pixel values in a binary document could introduce irregularities that are very visually noticeable. In this paper, a new method is proposed for object-based text document authentication, in which I propose a different approach where a text document is signed by shifting individual words slightly left or right from their original positions to make the center of gravity for each line fall in with the m
... Show MoreRecently a large number of extensive studies have amassed that describe the removal of dyes from water and wastewater using natural adsorbents and modified materials. Methyl orange dye is found in wastewater streams from various industries that include textiles, plastics, printing and paper among other sources. This article reviews methyl orange adsorption onto natural and modified materials. Despite many techniques available, adsorption stands out for efficient water and wastewater treatment for its ease of operation, flexibility and large-scale removal of colorants. It also has a significant potential for regeneration recovery and recycling of adsorbents in comparison to other water treatment methods. The adsorbents described herein were
... Show More