Data scarcity is a major challenge when training deep learning (DL) models. DL demands a large amount of data to achieve exceptional performance. Unfortunately, many applications have small or inadequate data to train DL frameworks. Usually, manual labeling is needed to provide labeled data, which typically involves human annotators with a vast background of knowledge. This annotation process is costly, time-consuming, and error-prone. Usually, every DL framework is fed by a significant amount of labeled data to automatically learn representations. Ultimately, a larger amount of data would generate a better DL model and its performance is also application dependent. This issue is the main barrier for many applications dismissing the use of DL. Having sufficient data is the first step toward any successful and trustworthy DL application. This paper presents a holistic survey on state-of-the-art techniques to deal with training DL models to overcome three challenges including small, imbalanced datasets, and lack of generalization. This survey starts by listing the learning techniques. Next, the types of DL architectures are introduced. After that, state-of-the-art solutions to address the issue of lack of training data are listed, such as Transfer Learning (TL), Self-Supervised Learning (SSL), Generative Adversarial Networks (GANs), Model Architecture (MA), Physics-Informed Neural Network (PINN), and Deep Synthetic Minority Oversampling Technique (DeepSMOTE). Then, these solutions were followed by some related tips about data acquisition needed prior to training purposes, as well as recommendations for ensuring the trustworthiness of the training dataset. The survey ends with a list of applications that suffer from data scarcity, several alternatives are proposed in order to generate more data in each application including Electromagnetic Imaging (EMI), Civil Structural Health Monitoring, Medical imaging, Meteorology, Wireless Communications, Fluid Mechanics, Microelectromechanical system, and Cybersecurity. To the best of the authors’ knowledge, this is the first review that offers a comprehensive overview on strategies to tackle data scarcity in DL.
The purpose of this research is to find the estimator of the average proportion of defectives based on attribute samples. That have been curtailed either with rejection of a lot finding the kth defective or with acceptance on finding the kth non defective.
The MLE (Maximum likelihood estimator) is derived. And also the ASN in Single Curtailed Sampling has been derived and we obtain a simplified Formula All the Notations needed are explained.
The research seeks to identify the image of foreign oil companies operating in Iraq among the public of Basra, and the research aims to clarify the mental image of foreign oil companies among the Iraqi public, and to identify the extent to which the Iraqi public benefit from the social responsibility programs offered by foreign oil companies and their contribution to improving the standard of living and services for the population. Nearby areas and society as a whole, the research is classified within descriptive research, and the researcher used the survey method for the Iraqi public in Basra governorate, which includes the areas in which these companies are located, and he used the scale tool to find out, so he distributed 600 que
... Show MoreThe key objective of the study is to understand the best processes that are currently used in managing talent in Australian higher education (AHE) and design a quantitative measurement of talent management processes (TMPs) for the higher education (HE) sector.
The three qualitative multi-method studies that are commonly used in empirical studies, namely, brainstorming, focus group discussions and semi-structured individual interviews were considered. Twenty
Two experiments were carried out, the first at the College of Agriculture - University of Baghdad during spring season 2017 Everest cv. class (Elite) was used to study the effect of foliar application of calcium and magnesium and addition of humic acid to the soil on potato growth and yield, The layout of the experiment was factorial within RCBD design using three replicates. Calcium and Magnesium sprayed with concentrations (0, 500, 1000 mg.L-1), while the humic acid was added to the soil with (0, 0.75 gm.m2), The second experiment included storage of tubers produced from the spring season, with to study the effect of field treatments on improving the storability of the tubers. The results showed that the treatment of calci
... Show MoreThe great importance of training made it as an investment for the organization, and assert the Quality of performance which support it by prepare the employee to the Current and future Jobs . The Research problem a rounded about How to measure the impact of training based on (ISO 10015) and its effect on the Quality of performance , How to evaluation the results of training to attained the training goals . The Research aims to find out the effects of application of international standard guidelines (ISO 10015) to attained the quality of audit work achieved in the Federal Board of Supreme Audit. The Research sought to achieve a number of objectives cognitive and applied on the basis of four key assumptions, and other su
... Show MoreIn this work, two different laser dye solutions were used to host highly-pure silicon nitride nanoparticles as scattering centers to fabricate random gain media. The laser dye was dissolved in three different solvents (ethanol, methanol and acetone) and the final results were obtained for methanol only. The silicon nitride nanoparticles were synthesized by dc reactive magnetron sputtering technique with average particle size of 35 nm. The random gain medium was made as a solid rod with high spectral efficiency and low production cost. Optical emission with narrow linewidth was detected at 532-534 nm as 9 mg of silicon nitride nanoparticles were added to the 10 -5 M dye solution. The FWHM of 0.3 and 3.52 nm was determined for Rhodamine B and
... Show MoreTor (The Onion Routing) network was designed to enable users to browse the Internet anonymously. It is known for its anonymity and privacy security feature against many agents who desire to observe the area of users or chase users’ browsing conventions. This anonymity stems from the encryption and decryption of Tor traffic. That is, the client’s traffic should be subject to encryption and decryption before the sending and receiving process, which leads to delay and even interruption in data flow. The exchange of cryptographic keys between network devices plays a pivotal and critical role in facilitating secure communication and ensuring the integrity of cryptographic procedures. This essential process is time-consuming, which causes del
... Show More