Clinical keratoconus (KCN) detection is a challenging and time-consuming task. In the diagnosis process, ophthalmologists must revise demographic and clinical ophthalmic examinations. The latter include slit-lamb, corneal topographic maps, and Pentacam indices (PI). We propose an Ensemble of Deep Transfer Learning (EDTL) based on corneal topographic maps. We consider four pretrained networks, SqueezeNet (SqN), AlexNet (AN), ShuffleNet (SfN), and MobileNet-v2 (MN), and fine-tune them on a dataset of KCN and normal cases, each including four topographic maps. We also consider a PI classifier. Then, our EDTL method combines the output probabilities of each of the five classifiers to obtain a decision based on the fusion of probabilities. Individually, the classifier based on PI achieved 93.1% accuracy, whereas the deep classifiers reached classification accuracies over 90% only in isolated cases. Overall, the average accuracy of the deep networks over the four corneal maps ranged from 86% (SfN) to 89.9% (AN). The classifier ensemble increased the accuracy of the deep classifiers based on corneal maps to values ranging (92.2% to 93.1%) for SqN and (93.1% to 94.8%) for AN. Including in the ensemble-specific combinations of corneal maps’ classifiers and PI increased the accuracy to 98.3%. Moreover, visualization of first learner filters in the networks and Grad-CAMs confirmed that the networks had learned relevant clinical features. This study shows the potential of creating ensembles of deep classifiers fine-tuned with a transfer learning strategy as it resulted in an improved accuracy while showing learnable filters and Grad-CAMs that agree with clinical knowledge. This is a step further towards the potential clinical deployment of an improved computer-assisted diagnosis system for KCN detection to help ophthalmologists to confirm the clinical decision and to perform fast and accurate KCN treatment.
Among the different passive techniques heat pipe heat exchanger (HPHE) seems to be the most effective one for energy saving in heating ventilation and air conditioning system (HVAC). The applications for nanofluids with high conductivity are favorable to increase the thermal performance in HPHE. Even though the nanofluid has the higher heat conduction coefficient that dispels more heat theoretically but the higher concentration will make clustering .Clustering is a problem that must be solved before nanofluids can be considered for long-term practical uses. Results showed that the maximum value of relative power is 0.13 mW at nanofluid compared with other concentrations due to the low density of nanofluid at this concentration. For highe
... Show Morein this article, we present a definition of k-generalized map independent of non-expansive map and give infinite families of non-expansive and k-generalized maps new iterative algorithms. Such algorithms are also studied in the Hilbert spaces as the potential to exist for asymptotic common fixed point.
In this research, a study is introduced on the effect of several environmental factors on the performance of an already constructed quality inspection system, which was designed using a transfer learning approach based on convolutional neural networks. The system comprised two sets of layers, transferred layers set from an already trained model (DenseNet121) and a custom classification layers set. It was designed to discriminate between damaged and undamaged helical gears according to the configuration of the gear regardless to its dimensions, and the model showed good performance discriminating between the two products at ideal conditions of high-resolution images. So, this study aimed at testing the system performance at poo
... Show MoreIn this research, a study is introduced on the effect of several environmental factors on the performance of an already constructed quality inspection system, which was designed using a transfer learning approach based on convolutional neural networks. The system comprised two sets of layers, transferred layers set from an already trained model (DenseNet121) and a custom classification layers set. It was designed to discriminate between damaged and undamaged helical gears according to the configuration of the gear regardless to its dimensions, and the model showed good performance discriminating between the two products at ideal conditions of high-resolution images.
So, this study aimed at testing the system performance at poor s
... Show MoreThere is a growing need for up-to-date data for rapid decision making in the modern digital age. Recently, the need for high-resolution topographic maps is highly demanding by most mapping clients. With the maturing automatic structure from mobile and multi-view stereoscopy software, small organizations and individuals now have the ability to make their own surveys based on mobile mapping devices. This study looks at how feasible using low-cost Unmanned Aerial Vehicle (UAV) as a mobile mapping device for photogrammetric topographical surveys. It is showing the impact of different UAV flight settings and parameters on the accuracy of mapping products. An automatic scenario for photogra
The Diffie-Hellman is a key exchange protocol to provide a way to transfer shared secret keys between two parties, although those parties might never have communicated together. This paper suggested a new way to transfer keys through public or non-secure channels depending on the sent video files over the channel and then extract keys. The proposed method of key generation depends on the video file content by using the entropy value of the video frames. The proposed system solves the weaknesses in the Diffie-Hellman key exchange algorithm, which is MIMA (Man-in-the-Middle attack) and DLA( Discrete logarithm attack). When the method used high definition videos with a vast amount of data, the keys generated with a large number up to 5
... Show MoreData scarcity is a major challenge when training deep learning (DL) models. DL demands a large amount of data to achieve exceptional performance. Unfortunately, many applications have small or inadequate data to train DL frameworks. Usually, manual labeling is needed to provide labeled data, which typically involves human annotators with a vast background of knowledge. This annotation process is costly, time-consuming, and error-prone. Usually, every DL framework is fed by a significant amount of labeled data to automatically learn representations. Ultimately, a larger amount of data would generate a better DL model and its performance is also application dependent. This issue is the main barrier for
A total of 37 Staphylococcus epidermidis isolates, isolated from corneal scraping of patients with bacterial keratitis and 20 isolates from healthy eyes (as control) (all isolates, isolated from, Ibn Al- Haietham eye hospital / Baghdad), were tested for slime production, 52.63% of all isolates were positive-slime production (23 isolates from patients and 7 isolates from controls). It was found that positive-slime producing S. epidermidis were exhibited a high resistance to antibiotics as compared to negative-slime producing isolates.
This work implements the face recognition system based on two stages, the first stage is feature extraction stage and the second stage is the classification stage. The feature extraction stage consists of Self-Organizing Maps (SOM) in a hierarchical format in conjunction with Gabor Filters and local image sampling. Different types of SOM’s were used and a comparison between the results from these SOM’s was given.
The next stage is the classification stage, and consists of self-organizing map neural network; the goal of this stage is to find the similar image to the input image. The proposal method algorithm implemented by using C++ packages, this work is successful classifier for a face database consist of 20
... Show MoreThe use of deep learning.