A substantial portion of today’s multimedia data exists in the form of unstructured text. However, the unstructured nature of text poses a significant task in meeting users’ information requirements. Text classification (TC) has been extensively employed in text mining to facilitate multimedia data processing. However, accurately categorizing texts becomes challenging due to the increasing presence of non-informative features within the corpus. Several reviews on TC, encompassing various feature selection (FS) approaches to eliminate non-informative features, have been previously published. However, these reviews do not adequately cover the recently explored approaches to TC problem-solving utilizing FS, such as optimization techniques. This study comprehensively analyzes different FS approaches based on optimization algorithms for TC. We begin by introducing the primary phases involved in implementing TC. Subsequently, we explore a wide range of FS approaches for categorizing text documents and attempt to organize the existing works into four fundamental approaches: filter, wrapper, hybrid, and embedded. Furthermore, we review four optimization algorithms utilized in solving text FS problems: swarm intelligence-based, evolutionary-based, physics-based, and human behavior-related algorithms. We discuss the advantages and disadvantages of state-of-the-art studies that employ optimization algorithms for text FS methods. Additionally, we consider several aspects of each proposed method and thoroughly discuss the challenges associated with datasets, FS approaches, optimization algorithms, machine learning classifiers, and evaluation criteria employed to assess new and existing techniques. Finally, by identifying research gaps and proposing future directions, our review provides valuable guidance to researchers in developing and situating further studies within the current body of literature.
This paper proposes a novel meta-heuristic optimization algorithm called the fine-tuning meta-heuristic algorithm (FTMA) for solving global optimization problems. In this algorithm, the solutions are fine-tuned using the fundamental steps in meta-heuristic optimization, namely, exploration, exploitation, and randomization, in such a way that if one step improves the solution, then it is unnecessary to execute the remaining steps. The performance of the proposed FTMA has been compared with that of five other optimization algorithms over ten benchmark test functions. Nine of them are well-known and already exist in the literature, while the tenth one is proposed by the authors and introduced in this article. One test trial was shown t
... Show MoreAmong a variety of approaches introduced in the literature to establish duality theory, Fenchel duality was of great importance in convex analysis and optimization. In this paper we establish some conditions to obtain classical strong Fenchel duality for evenly convex optimization problems defined in infinite dimensional spaces. The objective function of the primal problem is a family of (possible) infinite even convex functions. The strong duality conditions we present are based on the consideration of the epigraphs of the c-conjugate of the dual objective functions and the ε-c-subdifferential of the primal objective functions.
Researchers are increasingly using multimodal biometrics to strengthen the security of biometric applications. In this study, a strong multimodal human identification model was developed to address the growing problem of spoofing attacks in biometric security systems. Through the use of metaheuristic optimization methods, such as the Genetic Algorithm(GA), Ant Colony Optimization(ACO), and Particle Swarm Optimization (PSO) for feature selection, this unique model incorporates three biometric modalities: face, iris, and fingerprint. Image pre-processing, feature extraction, critical image feature selection, and multibiometric recognition are the four main steps in the workflow of the system. To determine its performance, the model wa
... Show MoreIn the present paper, the researcher attempts to shed some light on the objective behind inserting some Qur'anic verses by Al-Zahraa (Peace Be Upon Her) in her revered speech. Besides, it tries to investigate the hidden meaning of these verses and to study them in the light of pragmaticreferences. This task is supported by Books of Tafseer as well as the books that explained this speech to arrive at its intended meaning. It is possible say that this is astep towards studying speeches of 'Ahlul Bayt' (People of the Prophet's household) in terms of modern linguistic studies, as well as employing modern methods to explore the aesthetic values of these texts.
Developing a new adaptive satellite images classification technique, based on a new way of merging between regression line of best fit and new empirical conditions methods. They are supervised methods to recognize different land cover types on Al habbinya region. These methods should be stand on physical ground that represents the reflection of land surface features. The first method has separated the arid lands and plants. Empirical thresholds of different TM combination bands; TM3, TM4, and TM5 were studied in the second method, to detect and separate water regions (shallow, bottomless, and very bottomless). The Optimum Index Factor (OIF) is computed for these combination bands, which realized
... Show MoreCurrently, one of the topical areas of application of machine learning methods is the prediction of material characteristics. The aim of this work is to develop machine learning models for determining the rheological properties of polymers from experimental stress relaxation curves. The paper presents an overview of the main directions of metaheuristic approaches (local search, evolutionary algorithms) to solving combinatorial optimization problems. Metaheuristic algorithms for solving some important combinatorial optimization problems are described, with special emphasis on the construction of decision trees. A comparative analysis of algorithms for solving the regression problem in CatBoost Regressor has been carried out. The object of
... Show MoreIn this study, a traumatic spinal cord injury (TSCI) classification system is proposed using a convolutional neural network (CNN) technique with automatically learned features from electromyography (EMG) signals for a non-human primate (NHP) model. A comparison between the proposed classification system and a classical classification method (k-nearest neighbors, kNN) is also presented. Developing such an NHP model with a suitable assessment tool (i.e., classifier) is a crucial step in detecting the effect of TSCI using EMG, which is expected to be essential in the evaluation of the efficacy of new TSCI treatments. Intramuscular EMG data were collected from an agonist/antagonist tail muscle pair for the pre- and post-spinal cord lesi
... Show MoreAbstract
Uncertainty, the deeply-rooted fact that surrounding the investment environment, especially the stock market which just prices have taken a specific trend until they moved to another one for its up or down. This means that the volatility characteristic of financial market requires the rational investor an argument led towards the adoption of planned acts to gain greater benefit in the goal of wealth maximizing. There is no possibility to achieve this goal without the burden of uncertainty and the risk of systematic fluctuations of investment returns in the financial market after the facts of efficient diversification have pro
... Show MoreThe rise of edge-cloud continuum computing is a result of the growing significance of edge computing, which has become a complementary or substitute option for traditional cloud services. The convergence of networking and computers presents a notable challenge due to their distinct historical development. Task scheduling is a major challenge in the context of edge-cloud continuum computing. The selection of the execution location of tasks, is crucial in meeting the quality-of-service (QoS) requirements of applications. An efficient scheduling strategy for distributing workloads among virtual machines in the edge-cloud continuum data center is mandatory to ensure the fulfilment of QoS requirements for both customer and service provider. E
... Show More