Plagiarism is described as using someone else's ideas or work without their permission. Using lexical and semantic text similarity notions, this paper presents a plagiarism detection system for examining suspicious texts against available sources on the Web. The user can upload suspicious files in pdf or docx formats. The system will search three popular search engines for the source text (Google, Bing, and Yahoo) and try to identify the top five results for each search engine on the first retrieved page. The corpus is made up of the downloaded files and scraped web page text of the search engines' results. The corpus text and suspicious documents will then be encoded as vectors. For lexical plagiarism detection, the system will leverage Jaccard similarity and Term Frequency-Inverse Document Frequency (TFIDF) techniques, while for semantic plagiarism detection, Doc2Vec and Sentence Bidirectional Encoder Representations from Transformers (SBERT) intelligent text representation models will be used. Following that, the system compares the suspicious text to the corpus text. Finally, a generated plagiarism report will show the total plagiarism ratio, the plagiarism ratio from each source, and other details.
Today, problems of spatial data integration have been further complicated by the rapid development in communication technologies and the increasing amount of available data sources on the World Wide Web. Thus, web-based geospatial data sources can be managed by different communities and the data themselves can vary in respect to quality, coverage, and purpose. Integrating such multiple geospatial datasets remains a challenge for geospatial data consumers. This paper concentrates on the integration of geometric and classification schemes for official data, such as Ordnance Survey (OS) national mapping data, with volunteered geographic information (VGI) data, such as the data derived from the OpenStreetMap (OSM) project. Useful descriptions o
... Show MoreAmong the different passive techniques heat pipe heat exchanger (HPHE) seems to be the most effective one for energy saving in heating ventilation and air conditioning system (HVAC). The applications for nanofluids with high conductivity are favorable to increase the thermal performance in HPHE. Even though the nanofluid has the higher heat conduction coefficient that dispels more heat theoretically but the higher concentration will make clustering .Clustering is a problem that must be solved before nanofluids can be considered for long-term practical uses. Results showed that the maximum value of relative power is 0.13 mW at nanofluid compared with other concentrations due to the low density of nanofluid at this concentration. For highe
... Show MoreCutaneous Leishmaniasis (CL) is an endemic disease and one of the major health problems in Iraq. Leishmania tropica is known as the causative agent of Cutaneous Leishmaniasis in Baghdad.The classical serological methods of diagnosing leishmaniasis is a poor sensitivity especially for the sub genus and time consuming Here we have investigated two primer pairs, one specific for Leishmania as genus and the primer specific for the species of L. tropica to be detected by polymerase chain reaction (PCR).Samples were collected from (AL-karama Teaching Hospital) and whole genomic DNA was extracted from axenic promastigotes.The extracted DNA was amplified by PCRwith two KDNA primer pairs, for genus specific (13A/13B) and (Lmj4/Uni21) to identify
... Show MoreWorldwide, there is an increased reliance on COVID-19-related health messages to curb the COVID-19 outbreak. Therefore, it is vital to provide a well-prepared and authentic translation of English-language messages to reach culturally and linguistically diverse audiences. However, few studies, if any, focus on how non-English-speaking readers receive and linguistically accept the lexical choices in the messages translated into their language. The present study tested a sample of translated Arabic COVID-19-related texts that were obtained from the World Health Organization and Australian New South Wales Health websites. This study investigated to that extent Arabic readers would receive translated COVID-19 health messages and whether the t
... Show MoreEstimating the semantic similarity between short texts plays an increasingly prominent role in many fields related to text mining and natural language processing applications, especially with the large increase in the volume of textual data that is produced daily. Traditional approaches for calculating the degree of similarity between two texts, based on the words they share, do not perform well with short texts because two similar texts may be written in different terms by employing synonyms. As a result, short texts should be semantically compared. In this paper, a semantic similarity measurement method between texts is presented which combines knowledge-based and corpus-based semantic information to build a semantic network that repre
... Show MoreThe current study was designed to investigate the presence of aflatoxin M1 in 25 samples of pasteurized canned milk which collected randomly from some Iraqi local markets using ELISA technique. Aflatoxin M1 was present in 21 samples, the concentration of aflatoxin M1 ranged from (0.25-50 ppb). UV radiation (365nm wave length) was used for detoxification of aflatoxin M1 (sample with highest concentration /50 ppb of aflatoxin M1 in two different volumes ((25 & 50 ml)) for two different time (15 & 30 min) and 30, 60, 90 cm distance between lamp and milk layer were used for this purpose). Results showed that distance between lamp and milk layer was the most effective parameter in reduction of aflatoxin M1, and whenever the distance increase the
... Show MoreMany approaches of different complexity already exist to edge detection in
color images. Nevertheless, the question remains of how different are the results
when employing computational costly techniques instead of simple ones. This
paper presents a comparative study on two approaches to color edge detection to
reduce noise in image. The approaches are based on the Sobel operator and the
Laplace operator. Furthermore, an efficient algorithm for implementing the two
operators is presented. The operators have been applied to real images. The results
are presented in this paper. It is shown that the quality of the results increases by
using second derivative operator (Laplace operator). And noise reduced in a good
Heart disease identification is one of the most challenging task that requires highly experienced cardiologists. However, in developing nations such as Ethiopia, there are a few cardiologists and heart disease detection is more challenging. As an alternative solution to cardiologist, this study proposed a more effective model for heart disease detection by employing random forest and sequential feature selection (SFS). SFS is an effective approach to improve the performance of random forest model on heart disease detection. SFS removes unrelated features in heart disease dataset that tends to mislead random forest model on heart disease detection. Thus, removing inappropriate and duplicate features from the training set with sequential f
... Show More