Data scarcity is a major challenge when training deep learning (DL) models. DL demands a large amount of data to achieve exceptional performance. Unfortunately, many applications have small or inadequate data to train DL frameworks. Usually, manual labeling is needed to provide labeled data, which typically involves human annotators with a vast background of knowledge. This annotation process is costly, time-consuming, and error-prone. Usually, every DL framework is fed by a significant amount of labeled data to automatically learn representations. Ultimately, a larger amount of data would generate a better DL model and its performance is also application dependent. This issue is the main barrier for many applications dismissing the use of DL. Having sufficient data is the first step toward any successful and trustworthy DL application. This paper presents a holistic survey on state-of-the-art techniques to deal with training DL models to overcome three challenges including small, imbalanced datasets, and lack of generalization. This survey starts by listing the learning techniques. Next, the types of DL architectures are introduced. After that, state-of-the-art solutions to address the issue of lack of training data are listed, such as Transfer Learning (TL), Self-Supervised Learning (SSL), Generative Adversarial Networks (GANs), Model Architecture (MA), Physics-Informed Neural Network (PINN), and Deep Synthetic Minority Oversampling Technique (DeepSMOTE). Then, these solutions were followed by some related tips about data acquisition needed prior to training purposes, as well as recommendations for ensuring the trustworthiness of the training dataset. The survey ends with a list of applications that suffer from data scarcity, several alternatives are proposed in order to generate more data in each application including Electromagnetic Imaging (EMI), Civil Structural Health Monitoring, Medical imaging, Meteorology, Wireless Communications, Fluid Mechanics, Microelectromechanical system, and Cybersecurity. To the best of the authors’ knowledge, this is the first review that offers a comprehensive overview on strategies to tackle data scarcity in DL.
There is no access to basic sanitation for half the world's population, leading to Socioeconomic issues, such as scarcity of drinking water and the spread of diseases. In this way, it is of vital importance to develop water management technologies relevant to the target population. In addition, in the separation form of water treatment, the compound often used as a coagulant in water treatment is aluminum sulfate, which provides good results for raw water turbidity and color removal. Studies show, however, that its deposition in the human body, even Alzheimer's disease, can cause serious harm to health and disease development. The study aims to improve the coagulation/flocculation stage related to the amount of flakes, i
... Show MoreTime series analysis is the statistical approach used to analyze a series of data. Time series is the most popular statistical method for forecasting, which is widely used in several statistical and economic applications. The wavelet transform is a powerful mathematical technique that converts an analyzed signal into a time-frequency representation. The wavelet transform method provides signal information in both the time domain and frequency domain. The aims of this study are to propose a wavelet function by derivation of a quotient from two different Fibonacci coefficient polynomials, as well as a comparison between ARIMA and wavelet-ARIMA. The time series data for daily wind speed is used for this study. From the obtained results, the
... Show MoreThe research aims to identify the effect of the training program that is based on integrating futuristic thinking skills with classroom interaction patterns on mathematics teachers in order to provide their students with creative solution skills. The research sample consisted of 31teachers (15 teachers for the experimental group and 16 for the control groups). The researcher developed a measure for the academic self-efficacy consisting of (39) items. Its validity, reliability, coefficient of difficulty and discriminatory power were estimated. To analyze the findings, the researcher adopted the Mann-Whitney (U) test and the effect size, and the findings were as follows: There is a statistically significant difference at the significance leve
... Show MoreA thin film of AgInSe2 and Ag1-xCuxInSe2 as well as n-Ag1-xCuxInSe2 /p-Si heterojunction with different Cu ratios (0, 0.1, 0.2) has been successfully fabricated by thermal evaporation method as absorbent layer with thickness about 700 nm and ZnTe as window layer with thickness about 100 nm. We made a multi-layer of p-ZnTe/n-AgCuInSe2/p-Si structures, In the present work, the conversion efficiency (η) increased when added the Cu and when used p-ZnTe as a window layer (WL) the bandgap energy of the direct transition decreases from 1.75 eV (Cu=0.0) to 1.48 eV (Cu=0.2 nm) and the bandgap energy for ZnTe=2.35 eV. The measurements of the electrical properties for prepared films showed that the D.C electrical conductivity (σd.c) increase
... Show MoreIn the last two decades, networks had been changed according to the rapid changing in its requirements. The current Data Center Networks have large number of hosts (tens or thousands) with special needs of bandwidth as the cloud network and the multimedia content computing is increased. The conventional Data Center Networks (DCNs) are highlighted by the increased number of users and bandwidth requirements which in turn have many implementation limitations. The current networking devices with its control and forwarding planes coupling result in network architectures are not suitable for dynamic computing and storage needs. Software Defined networking (SDN) is introduced to change this notion of traditional networks by decoupling control and
... Show MoreThe Internet is providing vital communications between millions of individuals. It is also more and more utilized as one of the commerce tools; thus, security is of high importance for securing communications and protecting vital information. Cryptography algorithms are essential in the field of security. Brute force attacks are the major Data Encryption Standard attacks. This is the main reason that warranted the need to use the improved structure of the Data Encryption Standard algorithm. This paper proposes a new, improved structure for Data Encryption Standard to make it secure and immune to attacks. The improved structure of Data Encryption Standard was accomplished using standard Data Encryption Standard with a new way of two key gene
... Show MoreThis work presents an experimental study of heat transfer and flow of distilled water and metal oxide nanofluid Fe3O4-distilled water at concentrations of (φ = 0.3, 0.6, 0.9 %) by volume in a horizontal pipe with constant magnetic field. All the tests are carried out with Reynolds number range (2900-9820) and uniform heat flux (11262-19562 W/m2). The results show that, the nanofluid concentration and magnetic intensity increase, the Nusselt number increases. The maximum enhancement in Nusselt number with magnetic nanofluid is (5.4 %, 26.4 %, 42.7 %) for volume concentration (0.3, 0.6, 0.9 %) respectively. The enhancement is maximized with magnetic intensity (0.1, 0.2, 0.3 tesla) respectively to (43.9, 44
... Show More