Data scarcity is a major challenge when training deep learning (DL) models. DL demands a large amount of data to achieve exceptional performance. Unfortunately, many applications have small or inadequate data to train DL frameworks. Usually, manual labeling is needed to provide labeled data, which typically involves human annotators with a vast background of knowledge. This annotation process is costly, time-consuming, and error-prone. Usually, every DL framework is fed by a significant amount of labeled data to automatically learn representations. Ultimately, a larger amount of data would generate a better DL model and its performance is also application dependent. This issue is the main barrier for many applications dismissing the use of DL. Having sufficient data is the first step toward any successful and trustworthy DL application. This paper presents a holistic survey on state-of-the-art techniques to deal with training DL models to overcome three challenges including small, imbalanced datasets, and lack of generalization. This survey starts by listing the learning techniques. Next, the types of DL architectures are introduced. After that, state-of-the-art solutions to address the issue of lack of training data are listed, such as Transfer Learning (TL), Self-Supervised Learning (SSL), Generative Adversarial Networks (GANs), Model Architecture (MA), Physics-Informed Neural Network (PINN), and Deep Synthetic Minority Oversampling Technique (DeepSMOTE). Then, these solutions were followed by some related tips about data acquisition needed prior to training purposes, as well as recommendations for ensuring the trustworthiness of the training dataset. The survey ends with a list of applications that suffer from data scarcity, several alternatives are proposed in order to generate more data in each application including Electromagnetic Imaging (EMI), Civil Structural Health Monitoring, Medical imaging, Meteorology, Wireless Communications, Fluid Mechanics, Microelectromechanical system, and Cybersecurity. To the best of the authors’ knowledge, this is the first review that offers a comprehensive overview on strategies to tackle data scarcity in DL.
Mixed convection heat transfer in a vertical concentric annulus packed with a metallic porous media and heated at a constant heat flux is experimentally investigated with water as the working fluid. A series of experiments have been carried out with a Rayleigh number range from Ra=122418.92 to 372579.31 and Reynolds number that based on the particles diameter of Red=14.62, 19.48 and 24.36. Under steady state condition, the measured data were collected and analyzed. Results show that the wall surface temperatures are affected by the imposed heat flux variation and Reynolds number variation. The variation of the local heat transfer coefficient and the mean Nusselt number are presented and analyzed. An empirical
... Show MoreKnowledge of the distribution of the rock mechanical properties along the depth of the wells is an important task for many applications related to reservoir geomechanics. Such these applications are wellbore stability analysis, hydraulic fracturing, reservoir compaction and subsidence, sand production, and fault reactivation. A major challenge with determining the rock mechanical properties is that they are not directly measured at the wellbore. They can be only sampled at well location using rock testing. Furthermore, the core analysis provides discrete data measurements for specific depth as well as it is often available only for a few wells in a field of interest. This study presents a methodology to generate synthetic-geomechani
... Show MoreThe paper presents a highly accurate power flow solution, reducing the possibility of ending at local minima, by using Real-Coded Genetic Algorithm (RCGA) with system reduction and restoration. The proposed method (RCGA) is modified to reduce the total computing time by reducing the system in size to that of the generator buses, which, for any realistic system, will be smaller in number, and the load buses are eliminated. Then solving the power flow problem for the generator buses only by real-coded GA to calculate the voltage phase angles, whereas the voltage magnitudes are specified resulted in reduced computation time for the solution. Then the system is restored by calculating the voltages of the load buses in terms
... Show MoreCD40 is a type 1 transmembrane protein composed of 277 amino acids, and it belongs to the tumor necrosis factor receptor (TNFR) superfamily. It is expressed in a variety of cell types, including normal B cells, macrophages, dendritic cells, and endothelial cells, as a costimulatory molecule. This study aims to summarize the CD40 polymorphism effect and its susceptibility to immune-related disorders. The CD40 gene polymorphisms showed a significant association with different immune-related disorders and act as a risk factor for increased susceptibility to these diseases.
Age, hypertension, and diabetes can cause significant alterations in arterial structure and function, including changes in lumen diameter (LD), intimal-medial thickness (IMT), flow velocities, and arterial compliance. These are also considered risk markers of atherosclerosis and cerebrovascular disease. A difference between right and left carotid artery blood flow and IMT has been reported by some researchers, and a difference in the incidence of nonlacunar stroke has been reported between the right and left brain hemispheres. The aim of this study was to determine whether there are differences between the right and left common carotid arteries and internal carotid arteries in patient
Background: Economic Globalization affects work condition by increasing work stress. Chronic work stress ended with burnout syndrome.
Objectives: To estimate the prevalence of burnout syndrome and the association of job title, and violence with it among physicians in Baghdad, and to assess the burnout syndrome at patient and work levels by structured interviews.
Subjects and Methods: A cross section study was conducted on Physicians in Baghdad. Sampling was a multistage, stratified sampling to control the confounders in the design phase. A mixed qualitative and quantitative
... Show MoreIn this paper a system is designed and implemented using a Field Programmable Gate Array (FPGA) to move objects from a pick up location to a delivery location. This transportation of objects is done via a vehicle equipped with a robot arm and an FPGA. The path between the two locations is followed by recognizing a black line between them. The black line is sensed by Infrared sensors (IR) located on the front and on the back of the vehicle. The Robot was successfully implemented by programming the Field Programmable Gate Array with the designed system that was described as a state diagram and the robot operated properly.
Background: Economic Globalization affects work condition by increasing work stress. Chronic work stress ended with burnout syndrome. Objectives: To estimate the prevalence of burnout syndrome and the association of job title, and violence with it among physicians in Baghdad, and to assess the burnout syndrome at patient and work levels by structured interviews. Subjects and Methods: A cross section study was conducted on Physicians in Baghdad. Sampling was a multistage, stratified sampling to control the confounders in the design phase. A mixed qualitative and quantitative approach (triangulation) was used. Quantitative method used self-administered questionnaires of Maslach Burn out Inventory. Qualitative approach used an open-end
... Show More