Data scarcity is a major challenge when training deep learning (DL) models. DL demands a large amount of data to achieve exceptional performance. Unfortunately, many applications have small or inadequate data to train DL frameworks. Usually, manual labeling is needed to provide labeled data, which typically involves human annotators with a vast background of knowledge. This annotation process is costly, time-consuming, and error-prone. Usually, every DL framework is fed by a significant amount of labeled data to automatically learn representations. Ultimately, a larger amount of data would generate a better DL model and its performance is also application dependent. This issue is the main barrier for many applications dismissing the use of DL. Having sufficient data is the first step toward any successful and trustworthy DL application. This paper presents a holistic survey on state-of-the-art techniques to deal with training DL models to overcome three challenges including small, imbalanced datasets, and lack of generalization. This survey starts by listing the learning techniques. Next, the types of DL architectures are introduced. After that, state-of-the-art solutions to address the issue of lack of training data are listed, such as Transfer Learning (TL), Self-Supervised Learning (SSL), Generative Adversarial Networks (GANs), Model Architecture (MA), Physics-Informed Neural Network (PINN), and Deep Synthetic Minority Oversampling Technique (DeepSMOTE). Then, these solutions were followed by some related tips about data acquisition needed prior to training purposes, as well as recommendations for ensuring the trustworthiness of the training dataset. The survey ends with a list of applications that suffer from data scarcity, several alternatives are proposed in order to generate more data in each application including Electromagnetic Imaging (EMI), Civil Structural Health Monitoring, Medical imaging, Meteorology, Wireless Communications, Fluid Mechanics, Microelectromechanical system, and Cybersecurity. To the best of the authors’ knowledge, this is the first review that offers a comprehensive overview on strategies to tackle data scarcity in DL.
Products’ quality inspection is an important stage in every production route, in which the quality of the produced goods is estimated and compared with the desired specifications. With traditional inspection, the process rely on manual methods that generates various costs and large time consumption. On the contrary, today’s inspection systems that use modern techniques like computer vision, are more accurate and efficient. However, the amount of work needed to build a computer vision system based on classic techniques is relatively large, due to the issue of manually selecting and extracting features from digital images, which also produces labor costs for the system engineers. In this research, we pr
... Show MoreRecurrent strokes can be devastating, often resulting in severe disability or death. However, nearly 90% of the causes of recurrent stroke are modifiable, which means recurrent strokes can be averted by controlling risk factors, which are mainly behavioral and metabolic in nature. Thus, it shows that from the previous works that recurrent stroke prediction model could help in minimizing the possibility of getting recurrent stroke. Previous works have shown promising results in predicting first-time stroke cases with machine learning approaches. However, there are limited works on recurrent stroke prediction using machine learning methods. Hence, this work is proposed to perform an empirical analysis and to investigate machine learning al
... Show MorePublic-private partnership (PPP) has been used over the past 20-30 years by governments in developed countries to meet the public demand for infrastructural services. In Iraq, the PPP concept is comparatively new to the Government of Iraq (GoI), where the government has historically taken most of the responsibility for providing public services. There are few PPP projects in Iraq. However, the number is increasing. Recently the Iraqi market has experienced a number of attempts of PPP in different sectors, especially after the new investment law in 2006. The aim of this paper is to evaluate the investment environment in Iraq and to indicate the main factors affecting PPP in particular for infrastructure projects. Some literature review and
... Show MoreThe ability of the human brain to communicate with its environment has become a reality through the use of a Brain-Computer Interface (BCI)-based mechanism. Electroencephalography (EEG) has gained popularity as a non-invasive way of brain connection. Traditionally, the devices were used in clinical settings to detect various brain diseases. However, as technology advances, companies such as Emotiv and NeuroSky are developing low-cost, easily portable EEG-based consumer-grade devices that can be used in various application domains such as gaming, education. This article discusses the parts in which the EEG has been applied and how it has proven beneficial for those with severe motor disorders, rehabilitation, and as a form of communi
... Show MoreThe subject of this research involves studying adsorption to removal herbicide Atlantis WG from aqueous solutions by bentonite clay. The equilibrium concentration have been determined spectra photometry by using UV-Vis spectrophotometer. The experimental equilibrium sorption data were analyzed by two widely, Langmuir and Freundlish isotherm models. The Langmuir model gave a better fit than Freundlich model The adsorption amount of (Atlantis WG) increased when the temperature and pH decreased. The thermodynamic parameters like ?G, ?H, and ?S have been calculated from the effect of temperature on adsorption process, is exothermic. The kinetic of adsorption process was studied depending on Lagergren ,Morris ? Weber and Rauschenberg equati
... Show MoreActivated carbon prepared from date stones by chemical activation with ferric chloride (FAC) was used an adsorbent to remove phenolic compounds such as phenol (Ph) and p-nitro phenol (PNPh) from aqueous solutions. The influence of process variables represented by solution pH value (2-12), adsorbent to adsorbate weight ratio (0.2-1.8), and contact time (30-150 min) on removal percentage and adsorbed amount of Ph and PNPh onto FAC was studied. For PNPh adsorption,( 97.43 %) maximum removal percentage and (48.71 mg/g) adsorbed amount was achieved at (5) solution pH,( 1) adsorbent to adsorbate weight ratio, and (90 min) contact time. While for Ph adsorption, at (4) solution pH, (1.4) absorbent to adsorbate weight ratio, and (120 min) contact
... Show MoreThis study was conducted to determine the effect of vitamin A ( 10 mg/kg ) on avearage testis weight and sexual glands ( Prostate and Seminal Vesicle ) for albino male mice treated with Hexavalent chromium ( 1000 ppm ) .The current study 40 mice were divided into fife groups : 1st group treated with distilled water and considered an control group (C) / the 2nd group treated with sesame oil ( T1) / 3rd group was givin hexavalent chromium ( 1000 ppm ) (T2) / 4th group treated with vitamin A ( 10 mg / kg ) and exposed to hexavalent chromium ( 1000 ppm ) (T3) / 5th group treated with vitamin A ( 10 mg kg ) (T4) . The expermint lasted 35 day . the results showed a significant ( P ? 0.05 ) decrease in avearage testis weight and sexual glan
... Show MoreThis study investigates the impact of spatial resolution enhancement on supervised classification accuracy using Landsat 9 satellite imagery, achieved through pan-sharpening techniques leveraging Sentinel-2 data. Various methods were employed to synthesize a panchromatic (PAN) band from Sentinel-2 data, including dimension reduction algorithms and weighted averages based on correlation coefficients and standard deviation. Three pan-sharpening algorithms (Gram-Schmidt, Principal Components Analysis, Nearest Neighbour Diffusion) were employed, and their efficacy was assessed using seven fidelity criteria. Classification tasks were performed utilizing Support Vector Machine and Maximum Likelihood algorithms. Results reveal that specifi
... Show MoreTwitter data analysis is an emerging field of research that utilizes data collected from Twitter to address many issues such as disaster response, sentiment analysis, and demographic studies. The success of data analysis relies on collecting accurate and representative data of the studied group or phenomena to get the best results. Various twitter analysis applications rely on collecting the locations of the users sending the tweets, but this information is not always available. There are several attempts at estimating location based aspects of a tweet. However, there is a lack of attempts on investigating the data collection methods that are focused on location. In this paper, we investigate the two methods for obtaining location-based dat
... Show MoreThe main object of this article is to study and introduce a subclass of meromorphic univalent functions with fixed second positive defined by q-differed operator. Coefficient bounds, distortion and Growth theorems, and various are the obtained results.