Glaucoma is a visual disorder, which is one of the significant driving reason for visual impairment. Glaucoma leads to frustrate the visual information transmission to the brain. Dissimilar to other eye illness such as myopia and cataracts. The impact of glaucoma can’t be cured; The Disc Damage Likelihood Scale (DDLS) can be used to assess the Glaucoma. The proposed methodology suggested simple method to extract Neuroretinal rim (NRM) region then dividing the region into four sectors after that calculate the width for each sector and select the minimum value to use it in DDLS factor. The feature was fed to the SVM classification algorithm, the DDLS successfully classified Glaucoma disease with 70% percentage; moreover, when the dimensions of both Optic Disc(OD) and Optic Cup(OC) were used as additional features the accuracy rate raised to 91%.
Cutaneous Leishmaniasis (CL) is an endemic disease and one of the major health problems in Iraq. Leishmania tropica is known as the causative agent of Cutaneous Leishmaniasis in Baghdad.The classical serological methods of diagnosing leishmaniasis is a poor sensitivity especially for the sub genus and time consuming Here we have investigated two primer pairs, one specific for Leishmania as genus and the primer specific for the species of L. tropica to be detected by polymerase chain reaction (PCR).Samples were collected from (AL-karama Teaching Hospital) and whole genomic DNA was extracted from axenic promastigotes.The extracted DNA was amplified by PCRwith two KDNA primer pairs, for genus specific (13A/13B) and (Lmj4/Uni21) to identify
... Show MoreRecent research has shown that a Deoxyribonucleic Acid (DNA) has ability to be used to discover diseases in human body as its function can be used for an intrusion-detection system (IDS) to detect attacks against computer system and networks traffics. Three main factor influenced the accuracy of IDS based on DNA sequence, which is DNA encoding method, STR keys and classification method to classify the correctness of proposed method. The pioneer idea on attempt a DNA sequence for intrusion detection system is using a normal signature sequence with alignment threshold value, later used DNA encoding based cryptography, however the detection rate result is very low. Since the network traffic consists of 41 attributes, therefore we proposed the
... Show MoreBotnet detection develops a challenging problem in numerous fields such as order, cybersecurity, law, finance, healthcare, and so on. The botnet signifies the group of co-operated Internet connected devices controlled by cyber criminals for starting co-ordinated attacks and applying various malicious events. While the botnet is seamlessly dynamic with developing counter-measures projected by both network and host-based detection techniques, the convention techniques are failed to attain sufficient safety to botnet threats. Thus, machine learning approaches are established for detecting and classifying botnets for cybersecurity. This article presents a novel dragonfly algorithm with multi-class support vector machines enabled botnet
... Show MoreIn this paper, the botnet detection problem is defined as a feature selection problem and the genetic algorithm (GA) is used to search for the best significant combination of features from the entire search space of set of features. Furthermore, the Decision Tree (DT) classifier is used as an objective function to direct the ability of the proposed GA to locate the combination of features that can correctly classify the activities into normal traffics and botnet attacks. Two datasets namely the UNSW-NB15 and the Canadian Institute for Cybersecurity Intrusion Detection System 2017 (CICIDS2017), are used as evaluation datasets. The results reveal that the proposed DT-aware GA can effectively find the relevant
... Show MoreIn the task of detecting intrinsic plagiarism, the cases where reference corpus is absent are to be dealt with. This task is entirely based on inconsistencies within a given document. Detection of internal plagiarism has been considered as a classification problem. It can be estimated through taking into consideration self-based information from a given document.
The core contribution of the work proposed in this paper is associated with the document representation. Wherein, the document, also, the disjoint segments generated from it, have been represented as weight vectors demonstrating their main content. Where, for each element in these vectors, its average weight has been considered instead of its frequency.
Th
... Show MoreAn investigation was provided in this work for the host range of brown soft scale Coccus hesperidum Linnaeus in Baghdad Province. Five plant species were found infected by this insect, three of these species, Citrusaurantium L. (Rutaceae); Nerium oleander L. (Apocynaceae); Ficuscarica L. (Moraceae) reported earlier, and the remaining two, Dahlia pinnata Cav. (Asteraceae) and Myrtuscommunis L. (Myrtaceae) are recordedhere for the first time as host plants for this pest.
Abstract— The growing use of digital technologies across various sectors and daily activities has made handwriting recognition a popular research topic. Despite the continued relevance of handwriting, people still require the conversion of handwritten copies into digital versions that can be stored and shared digitally. Handwriting recognition involves the computer's strength to identify and understand legible handwriting input data from various sources, including document, photo-graphs and others. Handwriting recognition pose a complexity challenge due to the diversity in handwriting styles among different individuals especially in real time applications. In this paper, an automatic system was designed to handwriting recognition
... Show More