Decoding transient sEMG data for intent motion recognition in transhumeral amputees
...Show More Authors
The Internet is providing vital communications between millions of individuals. It is also more and more utilized as one of the commerce tools; thus, security is of high importance for securing communications and protecting vital information. Cryptography algorithms are essential in the field of security. Brute force attacks are the major Data Encryption Standard attacks. This is the main reason that warranted the need to use the improved structure of the Data Encryption Standard algorithm. This paper proposes a new, improved structure for Data Encryption Standard to make it secure and immune to attacks. The improved structure of Data Encryption Standard was accomplished using standard Data Encryption Standard with a new way of two key gene
... Show MoreSecure data communication across networks is always threatened with intrusion and abuse. Network Intrusion Detection System (IDS) is a valuable tool for in-depth defense of computer networks. Most research and applications in the field of intrusion detection systems was built based on analysing the several datasets that contain the attacks types using the classification of batch learning machine. The present study presents the intrusion detection system based on Data Stream Classification. Several data stream algorithms were applied on CICIDS2017 datasets which contain several new types of attacks. The results were evaluated to choose the best algorithm that satisfies high accuracy and low computation time.
The research aimed at measuring the compatibility of Big date with the organizational Ambidexterity dimensions of the Asia cell Mobile telecommunications company in Iraq in order to determine the possibility of adoption of Big data Triple as a approach to achieve organizational Ambidexterity.
The study adopted the descriptive analytical approach to collect and analyze the data collected by the questionnaire tool developed on the Likert scale After a comprehensive review of the literature related to the two basic study dimensions, the data has been subjected to many statistical treatments in accordance with res
... Show MoreThere is an assumption implicit but fundamental theory behind the decline by the time series used in the estimate, namely that the time series has a sleep feature Stationary or the language of Engle Gernger chains are integrated level zero, which indicated by I (0). It is well known, for example, tables of t-statistic is designed primarily to deal with the results of the regression that uses static strings. This assumption has been previously treated as an axiom the mid-seventies, where researchers are conducting studies of applied without taking into account the properties of time series used prior to the assessment, was to accept the results of these tests Bmanueh and delivery capabilities based on the applicability of the theo
... Show MoreAnomaly detection is still a difficult task. To address this problem, we propose to strengthen DBSCAN algorithm for the data by converting all data to the graph concept frame (CFG). As is well known that the work DBSCAN method used to compile the data set belong to the same species in a while it will be considered in the external behavior of the cluster as a noise or anomalies. It can detect anomalies by DBSCAN algorithm can detect abnormal points that are far from certain set threshold (extremism). However, the abnormalities are not those cases, abnormal and unusual or far from a specific group, There is a type of data that is do not happen repeatedly, but are considered abnormal for the group of known. The analysis showed DBSCAN using the
... Show MoreToday, the role of cloud computing in our day-to-day lives is very prominent. The cloud computing paradigm makes it possible to provide demand-based resources. Cloud computing has changed the way that organizations manage resources due to their robustness, low cost, and pervasive nature. Data security is usually realized using different methods such as encryption. However, the privacy of data is another important challenge that should be considered when transporting, storing, and analyzing data in the public cloud. In this paper, a new method is proposed to track malicious users who use their private key to decrypt data in a system, share it with others and cause system information leakage. Security policies are also considered to be int
... Show MoreMissing data is one of the problems that may occur in regression models. This problem is usually handled by deletion mechanism available in statistical software. This method reduces statistical inference values because deletion affects sample size. In this paper, Expectation Maximization algorithm (EM), Multicycle-Expectation-Conditional Maximization algorithm (MC-ECM), Expectation-Conditional Maximization Either (ECME), and Recurrent Neural Networks (RNN) are used to estimate multiple regression models when explanatory variables have some missing values. Experimental dataset were generated using Visual Basic programming language with missing values of explanatory variables according to a missing mechanism at random general pattern and s
... Show MoreAnemia is one of the common types of blood diseases, it lead to lack of number of RBCs (Red Blood Cell) and amount hemoglobin level in the blood is lower than normal.
In this paper a new algorithm is presented to recognize Anemia in digital images based on moment variant. The algorithm is accomplished using the following phases: preprocessing, segmentation, feature extraction and classification (using Decision Tree), the extracted features that are used for classification are Moment Invariant and Geometric Feature.
The Best obtained classification rates was 84% is obtained when using Moment Invariants features and 74 % is obtained when using Geometric Feature. Results indicate that the proposed algorithm is very effective in detect
There are many techniques for face recognition which compare the desired face image with a set of faces images stored in a database. Most of these techniques fail if faces images are exposed to high-density noise. Therefore, it is necessary to find a robust method to recognize the corrupted face image with a high density noise. In this work, face recognition algorithm was suggested by using the combination of de-noising filter and PCA. Many studies have shown that PCA has ability to solve the problem of noisy images and dimensionality reduction. However, in cases where faces images are exposed to high noise, the work of PCA in removing noise is useless, therefore adding a strong filter will help to im
... Show MoreMost systems are intelligent and the industrial world is moving now towards
technology. Most industrial systems are now computerized and offer a high speed.
However, Face recognition is a biometric system that can identify people from their
faces. For few number of people to be identified, it can be considered as a fast
system. When the number of people grew to be bigger, the system cannot be adopted
in a real-time application because its speed will degrade along with its accuracy.
However, the accuracy can be enhanced using pre-processing techniques but the
time delay is still a challenge. A series of experiments had been done on AT&TORL
database images using Enhanced Face Recognition System (EFRS) that is