Data scarcity is a major challenge when training deep learning (DL) models. DL demands a large amount of data to achieve exceptional performance. Unfortunately, many applications have small or inadequate data to train DL frameworks. Usually, manual labeling is needed to provide labeled data, which typically involves human annotators with a vast background of knowledge. This annotation process is costly, time-consuming, and error-prone. Usually, every DL framework is fed by a significant amount of labeled data to automatically learn representations. Ultimately, a larger amount of data would generate a better DL model and its performance is also application dependent. This issue is the main barrier for many applications dismissing the use of DL. Having sufficient data is the first step toward any successful and trustworthy DL application. This paper presents a holistic survey on state-of-the-art techniques to deal with training DL models to overcome three challenges including small, imbalanced datasets, and lack of generalization. This survey starts by listing the learning techniques. Next, the types of DL architectures are introduced. After that, state-of-the-art solutions to address the issue of lack of training data are listed, such as Transfer Learning (TL), Self-Supervised Learning (SSL), Generative Adversarial Networks (GANs), Model Architecture (MA), Physics-Informed Neural Network (PINN), and Deep Synthetic Minority Oversampling Technique (DeepSMOTE). Then, these solutions were followed by some related tips about data acquisition needed prior to training purposes, as well as recommendations for ensuring the trustworthiness of the training dataset. The survey ends with a list of applications that suffer from data scarcity, several alternatives are proposed in order to generate more data in each application including Electromagnetic Imaging (EMI), Civil Structural Health Monitoring, Medical imaging, Meteorology, Wireless Communications, Fluid Mechanics, Microelectromechanical system, and Cybersecurity. To the best of the authors’ knowledge, this is the first review that offers a comprehensive overview on strategies to tackle data scarcity in DL.
This study aimed to evaluate oral health (OH) and periodontal diseases (PD) awareness in the Iraqi population.
This study was a questionnaire‐based online survey of two weeks duration. The questionnaire was built using a Google platform and was distributed randomly via social media (Facebook and Telegram). The questionnaire consisted of a demographic data section and two other main sections for the evaluation of OH and PD awareness. Each response was marked with “1” for a positive answer and “0” for the other answers. For each respondent, answers were summed to give
This paper describes the use of microcomputer as a laboratory instrument system. The system is focused on three weather variables measurement, are temperature, wind speed, and wind direction. This instrument is a type of data acquisition system; in this paper we deal with the design and implementation of data acquisition system based on personal computer (Pentium) using Industry Standard Architecture (ISA)bus. The design of this system involves mainly a hardware implementation, and the software programs that are used for testing, measuring and control. The system can be used to display the required information that can be transferred and processed from the external field to the system. A visual basic language with Microsoft foundation cl
... Show MoreThe achievements of the art that we know today are questioned in motives that differ from what art knew before, including dramatic artistic transformations, which he called modern art.
In view of the enormity of such a topic, its ramifications and its complexity, it was necessary to confine its subject to the origin of the motives of the transformations of its first pioneers, and then to stand on what resulted from that of the data of vision in composition and drawing exclusively, and through exploration in that, we got to know the vitality of change from the art of its time.
And by examining the ruling contemporary philosophical concepts and their new standards and their epistemological role in contemporary life, since they includ
This paper presents a hybrid approach for solving null values problem; it hybridizes rough set theory with intelligent swarm algorithm. The proposed approach is a supervised learning model. A large set of complete data called learning data is used to find the decision rule sets that then have been used in solving the incomplete data problem. The intelligent swarm algorithm is used for feature selection which represents bees algorithm as heuristic search algorithm combined with rough set theory as evaluation function. Also another feature selection algorithm called ID3 is presented, it works as statistical algorithm instead of intelligent algorithm. A comparison between those two approaches is made in their performance for null values estima
... Show MoreThis article studied some linear and nonlinear optical characteristics of different pH solutions from anthocyanin dye extract at 180 oC from red cabbage. First, the linear spectral characteristics, including absorption and transmittance in the range 400-800 nm for anthocyanin solution 5% v/v with different pHs, were achieved utilizing a UV/VIS spectrophotometer. The experimental results reveal a shift in the absorption toward the longer wavelength direction as pH values increment. Then, the nonlinear features were measured using the Z-scan technique with a CW 532 nm laser to measure the nonlinear absorption coefficient through an open aperture. A close aperture (diameter 2 mm) calculates the nonlinear refractive index. The open Z-scan sh
... Show MoreThe adsorption of Cr (VI) from aqueous solution by spent tea leaves (STL) was studied at different initial Cr (VI) concentrations, adsorbent dose, pH and contact time under batch isotherm experiments The adsorption experiments were carried out at 30°C and the effects of the four parameters on chromium uptake to establish a mathematical model description percentage removal of Cr (VI). The
analysis results showed that the experimental data were adequately fitted to second order polynomial model with correlation coefficients for this model was (R2 = 0.9891). The optimum operating parameters of initial Cr (VI) concentrations, adsorbent dose, pH and contact time were 50 mg/l, 0.7625 g, 3 and 100 min, respectively. At these conditions, th
In this study, Yogurt was dried and milled, then shaked with distilled water to remove the soluble materials, then again dried and milled. Batch experiments were carried out to remove hexavalent chromium from aqueous solutions. Different parameters were optimized such as amount of adsorbent, treatment time, pH and concentration of adsorbate. The concentrations of Cr6+ in solutions are determined by UV-Visible spectrophotometer. Maximum percentage removal of Cr6+ was 82% at pH 2. Two equilibrium adsorption isotherms mechanisms are tested Langmuir and Freundlich, the results showed that the isotherm obeyed to Freundlich isotherm. Kinetic models were applied to the adsorption of Cr6+ ions on the adsorbents, ps
... Show More