Learning is the process of gaining knowledge and implementing this knowledge on behavior. The concept of learning is not strict to just human being, it expanded to include machine also. Now the machines can behave based on the gained knowledge learned from the environment. The learning process is evolving in both human and machine, to keep up with the technology in the world, the human learning evolved into micro-learning and the machine learning evolved to deep learning. In this paper, the evolution of learning is discussed as a formal survey accomplished with the foundation of machine learning and its evolved version of learning which is deep learning and micro-learning as a new learning technology can be implemented on human and machine learning. A procedural comparison is achieved to declare the purpose of this survey, also a related discussion integrates the aim of this study. Finally a concluded points are illustrated as outcome which summarized the practical evolution intervals of the machine learning different concepts.
The novel coronavirus 2019 (COVID-19) is a respiratory syndrome with similar traits to common pneumonia. This major pandemic has affected nations both socially and economically, disturbing everyday life and urging the scientific community to develop solutions for the diagnosis and prevention of COVID-19. Reverse transcriptase-polymerase chain reaction (RT–PCR) is the conventional approach used for detecting COVID-19. Nevertheless, the initial stage of the infection is less predictable in PCR tests, making early prediction challenging. A robust and alternative diagnostic method based on digital computerised technologies to support conventional methods would greatly help society. Therefore, this paper reviews recent research bas
... Show MoreText categorization refers to the process of grouping text or documents into classes or categories according to their content. Text categorization process consists of three phases which are: preprocessing, feature extraction and classification. In comparison to the English language, just few studies have been done to categorize and classify the Arabic language. For a variety of applications, such as text classification and clustering, Arabic text representation is a difficult task because Arabic language is noted for its richness, diversity, and complicated morphology. This paper presents a comprehensive analysis and a comparison for researchers in the last five years based on the dataset, year, algorithms and the accu
... Show MoreDistributed Denial of Service (DDoS) attacks on Web-based services have grown in both number and sophistication with the rise of advanced wireless technology and modern computing paradigms. Detecting these attacks in the sea of communication packets is very important. There were a lot of DDoS attacks that were directed at the network and transport layers at first. During the past few years, attackers have changed their strategies to try to get into the application layer. The application layer attacks could be more harmful and stealthier because the attack traffic and the normal traffic flows cannot be told apart. Distributed attacks are hard to fight because they can affect real computing resources as well as network bandwidth. DDoS attacks
... Show MoreThe Arabic Language is the native tongue of more than 400 million people around the world, it is also a language that carries an important religious and international weight. The Arabic language has taken its share of the huge technological explosion that has swept the world, and therefore it needs to be addressed with natural language processing applications and tasks.
This paper aims to survey and gather the most recent research related to Arabic Part of Speech (APoS), pointing to tagger methods used for the Arabic language, which ought to aim to constructing corpus for Arabic tongue. Many AI investigators and researchers have worked and performed POS utilizing various machine-learning methods, such as Hidden-Mark
... Show MoreText categorization refers to the process of grouping text or documents into classes or categories according to their content. Text categorization process consists of three phases which are: preprocessing, feature extraction and classification. In comparison to the English language, just few studies have been done to categorize and classify the Arabic language. For a variety of applications, such as text classification and clustering, Arabic text representation is a difficult task because Arabic language is noted for its richness, diversity, and complicated morphology. This paper presents a comprehensive analysis and a comparison for researchers in the last five years based on the dataset, year, algorithms and the accuracy th
... Show MoreArtificial intelligence techniques are reaching us in several forms, some of which are useful but can be exploited in a way that harms us. One of these forms is called deepfakes. Deepfakes is used to completely modify video (or image) content to display something that was not in it originally. The danger of deepfake technology impact on society through the loss of confidence in everything is published. Therefore, in this paper, we focus on deepfakedetection technology from the view of two concepts which are deep learning and forensic tools. The purpose of this survey is to give the reader a deeper overview of i) the environment of deepfake creation and detection, ii) how deep learning and forensic tools contributed to the detection
... Show MoreEmotion could be expressed through unimodal social behaviour’s or bimodal or it could be expressed through multimodal. This survey describes the background of facial emotion recognition and surveys the emotion recognition using visual modality. Some publicly available datasets are covered for performance evaluation. A summary of some of the research efforts to classify emotion using visual modality for the last five years from 2013 to 2018 is given in a tabular form.
Semantic segmentation realization and understanding is a stringent task not just for computer vision but also in the researches of the sciences of earth, semantic segmentation decompose compound architectures in one elements, the most mutual object in a civil outside or inside senses must classified then reinforced with information meaning of all object, it’s a method for labeling and clustering point cloud automatically. Three dimensions natural scenes classification need a point cloud dataset to representation data format as input, many challenge appeared with working of 3d data like: little number, resolution and accurate of three Dimensional dataset . Deep learning now is the po
Data scarcity is a major challenge when training deep learning (DL) models. DL demands a large amount of data to achieve exceptional performance. Unfortunately, many applications have small or inadequate data to train DL frameworks. Usually, manual labeling is needed to provide labeled data, which typically involves human annotators with a vast background of knowledge. This annotation process is costly, time-consuming, and error-prone. Usually, every DL framework is fed by a significant amount of labeled data to automatically learn representations. Ultimately, a larger amount of data would generate a better DL model and its performance is also application dependent. This issue is the main barrier for