This paper proposes a better solution for EEG-based brain language signals classification, it is using machine learning and optimization algorithms. This project aims to replace the brain signal classification for language processing tasks by achieving the higher accuracy and speed process. Features extraction is performed using a modified Discrete Wavelet Transform (DWT) in this study which increases the capability of capturing signal characteristics appropriately by decomposing EEG signals into significant frequency components. A Gray Wolf Optimization (GWO) algorithm method is applied to improve the results and select the optimal features which achieves more accurate results by selecting impactful features with maximum relevance while minimizing redundancy. This optimization process improves the performance of the classification model in general. In case of classification, the Support Vector Machine (SVM) and Neural Network (NN) hybrid model is presented. This combines an SVM classifier's capacity to manage functions in high dimensional space, as well as a neural network capacity to learn non-linearly with its feature (pattern learning). The model was trained and tested on an EEG dataset and performed a classification accuracy of 97%, indicating the robustness and efficacy of our method. The results indicate that this improved classifier is able to be used in brain–computer interface systems and neurologic evaluations. The combination of machine learning and optimization techniques has established this paradigm as a highly effective way to pursue further research in EEG signal processing for brain language recognition.
The presented study investigated the scheduling regarding jobs on a single machine. Each job will be processed with no interruptions and becomes available for the processing at time 0. The aim is finding a processing order with regard to jobs, minimizing total completion time , total late work , and maximal tardiness which is an NP-hard problem. In the theoretical part of the present work, the mathematical formula for the examined problem will be presented, and a sub-problem of the original problem of minimizing the multi-objective functions is introduced. Also, then the importance regarding the dominance rule (DR) that could be applied to the problem to improve good solutions will be shown. While in the practical part, two
... Show MoreApple slice grading is useful in post-harvest operations for sorting, grading, packaging, labeling, processing, storage, transportation, and meeting market demand and consumer preferences. Proper grading of apple slices can help ensure the quality, safety, and marketability of the final products, contributing to the post-harvest operations of the overall success of the apple industry. The article aims to create a convolutional neural network (CNN) model to classify images of apple slices after immersing them in atmospheric plasma at two different pressures (1 and 5 atm) and two different immersion times (3 and again 6 min) once and in filtered water based on the hardness of the slices usin
The background subtraction is a leading technique adopted for detecting the moving objects in video surveillance systems. Various background subtraction models have been applied to tackle different challenges in many surveillance environments. In this paper, we propose a model of pixel-based color-histogram and Fuzzy C-means (FCM) to obtain the background model using cosine similarity (CS) to measure the closeness between the current pixel and the background model and eventually determine the background and foreground pixel according to a tuned threshold. The performance of this model is benchmarked on CDnet2014 dynamic scenes dataset using statistical metrics. The results show a better performance against the state-of the art
... Show MoreBootstrap is one of an important re-sampling technique which has given the attention of researches recently. The presence of outliers in the original data set may cause serious problem to the classical bootstrap when the percentage of outliers are higher than the original one. Many methods are proposed to overcome this problem such Dynamic Robust Bootstrap for LTS (DRBLTS) and Weighted Bootstrap with Probability (WBP). This paper try to show the accuracy of parameters estimation by comparison the results of both methods. The bias , MSE and RMSE are considered. The criterion of the accuracy is based on the RMSE value since the method that provide us RMSE value smaller than other is con
... Show MoreAbstract—The upper limb amputation exerts a significant burden on the amputee, limiting their ability to perform everyday activities, and degrading their quality of life. Amputee patients’ quality of life can be improved if they have natural control over their prosthetic hands. Among the biological signals, most commonly used to predict upper limb motor intentions, surface electromyography (sEMG), and axial acceleration sensor signals are essential components of shoulder-level upper limb prosthetic hand control systems. In this work, a pattern recognition system is proposed to create a plan for categorizing high-level upper limb prostheses in seven various types of shoulder girdle motions. Thus, combining seven feature groups, w
... Show MoreIn this paper, the construction of Hermite wavelets functions and their operational matrix of integration is presented. The Hermite wavelets method is applied to solve nth order Volterra integro diferential equations (VIDE) by expanding the unknown functions, as series in terms of Hermite wavelets with unknown coefficients. Finally, two examples are given
Cryptography is the process of transforming message to avoid an unauthorized access of data. One of the main problems and an important part in cryptography with secret key algorithms is key. For higher level of secure communication key plays an important role. For increasing the level of security in any communication, both parties must have a copy of the secret key which, unfortunately, is not that easy to achieve. Triple Data Encryption Standard algorithm is weak due to its weak key generation, so that key must be reconfigured to make this algorithm more secure, effective, and strong. Encryption key enhances the Triple Data Encryption Standard algorithm securities. This paper proposed a combination of two efficient encryption algorithms to
... Show MoreAnomaly detection is still a difficult task. To address this problem, we propose to strengthen DBSCAN algorithm for the data by converting all data to the graph concept frame (CFG). As is well known that the work DBSCAN method used to compile the data set belong to the same species in a while it will be considered in the external behavior of the cluster as a noise or anomalies. It can detect anomalies by DBSCAN algorithm can detect abnormal points that are far from certain set threshold (extremism). However, the abnormalities are not those cases, abnormal and unusual or far from a specific group, There is a type of data that is do not happen repeatedly, but are considered abnormal for the group of known. The analysis showed DBSCAN using the
... Show MoreA two time step stochastic multi-variables multi-sites hydrological data forecasting model was developed and verified using a case study. The philosophy of this model is to use the cross-variables correlations, cross-sites correlations and the two steps time lag correlations simultaneously, for estimating the parameters of the model which then are modified using the mutation process of the genetic algorithm optimization model. The objective function that to be minimized is the Akiake test value. The case study is of four variables and three sites. The variables are the monthly air temperature, humidity, precipitation, and evaporation; the sites are Sulaimania, Chwarta, and Penjwin, which are located north Iraq. The model performance was
... Show MoreVariable selection is an essential and necessary task in the statistical modeling field. Several studies have triedto develop and standardize the process of variable selection, but it isdifficultto do so. The first question a researcher needs to ask himself/herself what are the most significant variables that should be used to describe a given dataset’s response. In thispaper, a new method for variable selection using Gibbs sampler techniqueshas beendeveloped.First, the model is defined, and the posterior distributions for all the parameters are derived.The new variable selection methodis tested usingfour simulation datasets. The new approachiscompared with some existingtechniques: Ordinary Least Squared (OLS), Least Absolute Shrinkage
... Show More