Classification of imbalanced data is an important issue. Many algorithms have been developed for classification, such as Back Propagation (BP) neural networks, decision tree, Bayesian networks etc., and have been used repeatedly in many fields. These algorithms speak of the problem of imbalanced data, where there are situations that belong to more classes than others. Imbalanced data result in poor performance and bias to a class without other classes. In this paper, we proposed three techniques based on the Over-Sampling (O.S.) technique for processing imbalanced dataset and redistributing it and converting it into balanced dataset. These techniques are (Improved Synthetic Minority Over-Sampling Technique (Improved SMOTE), Borderline-SMOTE + Imbalanced Ratio(IR), Adaptive Synthetic Sampling (ADASYN) +IR) Algorithm, where the work these techniques are generate the synthetic samples for the minority class to achieve balance between minority and majority classes and then calculate the IR between classes of minority and majority. Experimental results show ImprovedSMOTE algorithm outperform the Borderline-SMOTE + IR and ADASYN + IR algorithms because it achieves a high balance between minority and majority classes.
We have presented the distribution of the exponentiated expanded power function (EEPF) with four parameters, where this distribution was created by the exponentiated expanded method created by the scientist Gupta to expand the exponential distribution by adding a new shape parameter to the cumulative function of the distribution, resulting in a new distribution, and this method is characterized by obtaining a distribution that belongs for the exponential family. We also obtained a function of survival rate and failure rate for this distribution, where some mathematical properties were derived, then we used the method of maximum likelihood (ML) and method least squares developed (LSD)
... Show More
Abstract:
We can notice cluster data in social, health and behavioral sciences, so this type of data have a link between its observations and we can express these clusters through the relationship between measurements on units within the same group.
In this research, I estimate the reliability function of cluster function by using the seemingly unrelate
... Show MoreThis research aims to know the effectiveness of teaching with a proposed strategy according to the common Knowledge construction modelin mathematical proficiency among students of the second middle class. The researchers adopted the method of the experimental approach, as the experimental design was used for two independent and equal groups with a post-test. The experiment was applied to a sample consisting of (83) students divided into two groups: an experimental comprising (42) students and a control group, the second comprising (41) students., from Badr Shaker Al-Sayyab Intermediate School for Boys, for the first semester of the academic year (2021-2022), the two groups were rewarded in four variables: (chronological age calculated in mo
... Show MoreBackground:Â Various fluids in the oral environment can affect the surface roughness of resin composites. This in vitro study was conducted to determine the influence of the mouth rinses on surface roughness of two methacrylate-based resin (nanofilled and packable composite) and siloraine-based resin composites.
Materials and methods: Disc-shaped specimens (12 mm in diameter and 2mm in height) were prepared from three types of composi
... Show MoreIn this paper, we will study non parametric model when the response variable have missing data (non response) in observations it under missing mechanisms MCAR, then we suggest Kernel-Based Non-Parametric Single-Imputation instead of missing value and compare it with Nearest Neighbor Imputation by using the simulation about some difference models and with difference cases as the sample size, variance and rate of missing data.
The Weibull distribution is considered one of the Type-I Generalized Extreme Value (GEV) distribution, and it plays a crucial role in modeling extreme events in various fields, such as hydrology, finance, and environmental sciences. Bayesian methods play a strong, decisive role in estimating the parameters of the GEV distribution due to their ability to incorporate prior knowledge and handle small sample sizes effectively. In this research, we compare several shrinkage Bayesian estimation methods based on the squared error and the linear exponential loss functions. They were adopted and compared by the Monte Carlo simulation method. The performance of these methods is assessed based on their accuracy and computational efficiency in estimati
... Show MoreWe propose two simple, rapid, and convenient spectrophotometric methods which are described for the determination of cephalexin in bulk and its pharmaceutical preparations. They are based on the measurement of the flame atomic emission of potassium ion (in the first method) and colorimetric determination of the green colored solution at 610 nm formed after the reaction of cephalexin with potassium permanganate as an oxidant agent (in the second method) in basic medium. The working conditions of the methods are investigated and optimized. Beer's law plot shows a good correlation in the concentration range of 5-40?g ml-1. The detection limits are 2.573,2.814 ?g ml-1 for the flame emission photometric method and 1.844,2.016 ?g ml-1 for colo
... Show MoreThe permeability is the most important parameter that indicates how efficient the reservoir fluids flow through the rock pores to the wellbore. Well-log evaluation and core measurements techniques are typically used to estimate it. In this paper, the permeability has been predicted by using classical and Flow zone indicator methods. A comparison between the two methods shows the superiority of the FZI method correlations, these correlations can be used to estimate permeability in un-cored wells with a good approximation.