The deep learning algorithm has recently achieved a lot of success, especially in the field of computer vision. This research aims to describe the classification method applied to the dataset of multiple types of images (Synthetic Aperture Radar (SAR) images and non-SAR images). In such a classification, transfer learning was used followed by fine-tuning methods. Besides, pre-trained architectures were used on the known image database ImageNet. The model VGG16 was indeed used as a feature extractor and a new classifier was trained based on extracted features.The input data mainly focused on the dataset consist of five classes including the SAR images class (houses) and the non-SAR images classes (Cats, Dogs, Horses, and Humans). The Convolutional Neural Network (CNN) has been chosen as a better option for the training process because it produces a high accuracy. The final accuracy has reached 91.18% in five different classes. The results are discussed in terms of the probability of accuracy for each class in the image classification in percentage. Cats class got 99.6 %, while houses class got 100 %.Other types of classes were with an average score of 90 % and above.
Diagnosing heart disease has become a very important topic for researchers specializing in artificial intelligence, because intelligence is involved in most diseases, especially after the Corona pandemic, which forced the world to turn to intelligence. Therefore, the basic idea in this research was to shed light on the diagnosis of heart diseases by relying on deep learning of a pre-trained model (Efficient b3) under the premise of using the electrical signals of the electrocardiogram and resample the signal in order to introduce it to the neural network with only trimming processing operations because it is an electrical signal whose parameters cannot be changed. The data set (China Physiological Signal Challenge -cspsc2018) was ad
... Show MoreVariable selection is an essential and necessary task in the statistical modeling field. Several studies have triedto develop and standardize the process of variable selection, but it isdifficultto do so. The first question a researcher needs to ask himself/herself what are the most significant variables that should be used to describe a given dataset’s response. In thispaper, a new method for variable selection using Gibbs sampler techniqueshas beendeveloped.First, the model is defined, and the posterior distributions for all the parameters are derived.The new variable selection methodis tested usingfour simulation datasets. The new approachiscompared with some existingtechniques: Ordinary Least Squared (OLS), Least Absolute Shrinkage
... Show MoreAbstract
The study aims to examine the relationships between cognitive absorption and E-Learning readiness in the preparatory stage. The study sample consisted of (190) students who were chosen randomly. The Researcher has developed the cognitive absorption and E-Learning readiness scales. A correlational descriptive approach was adopted. The research revealed that there is a positive statistical relationship between cognitive absorption and eLearning readiness.
In data mining, classification is a form of data analysis that can be used to extract models describing important data classes. Two of the well known algorithms used in data mining classification are Backpropagation Neural Network (BNN) and Naïve Bayesian (NB). This paper investigates the performance of these two classification methods using the Car Evaluation dataset. Two models were built for both algorithms and the results were compared. Our experimental results indicated that the BNN classifier yield higher accuracy as compared to the NB classifier but it is less efficient because it is time-consuming and difficult to analyze due to its black-box implementation.
This research introduce a study with application on Principal Component Regression obtained from some of the explainatory variables to limitate Multicollinearity problem among these variables and gain staibilty in their estimations more than those which yield from Ordinary Least Squares. But the cost that we pay in the other hand losing a little power of the estimation of the predictive regression function in explaining the essential variations. A suggested numerical formula has been proposed and applied by the researchers as optimal solution, and vererifing the its efficiency by a program written by the researchers themselves for this porpuse through some creterions: Cumulative Percentage Variance, Coefficient of Determination, Variance
... Show MoreThis paper describes a number of new interleaving strategies based on the golden section. The new interleavers are called golden relative prime interleavers, golden interleavers, and dithered golden interleavers. The latter two approaches involve sorting a real-valued vector derived from the golden section. Random and so-called “spread” interleavers are also considered. Turbo-code performance results are presented and compared for the various interleaving strategies. Of the interleavers considered, the dithered golden interleaver typically provides the best performance, especially for low code rates and large block sizes. The golden relative prime interleaver is shown to work surprisingly well for high puncture rates. These interleav
... Show MoreIt is found that hypersensitive teeth have a larger number and wider patent tubules than those of non-sensitive teeth. Objective: The aim of this study is to compare between the effects of diode laser at different power densities, with and without sodium fluoride on the sealing of exposed dentinal tubules and dentin permeability. Materials and methods: 118 teeth were used. Samples were divided into three major groups. The first consisted of 100 teeth used for permeability test. The second consisted of 16 teeth for measuring external surface temperature elevation while irradiation. The third, in turn, consisted of one pair of teeth observed under SEM for dentine surface morphology analysis. Results: For dentin permeability measurement, th
... Show MoreThis paper proposes a new method Object Detection in Skin Cancer Image, the minimum
spanning tree Detection descriptor (MST). This ObjectDetection descriptor builds on the
structure of the minimum spanning tree constructed on the targettraining set of Skin Cancer
Images only. The Skin Cancer Image Detection of test objects relies on their distances to the
closest edge of thattree. Our experimentsshow that the Minimum Spanning Tree (MST) performs
especially well in case of Fogginessimage problems and in highNoisespaces for Skin Cancer
Image.
The proposed method of Object Detection Skin Cancer Image wasimplemented and tested on
different Skin Cancer Images. We obtained very good results . The experiment showed that