Iris research is focused on developing techniques for identifying and locating relevant biometric features, accurate segmentation and efficient computation while lending themselves to compression methods. Most iris segmentation methods are based on complex modelling of traits and characteristics which, in turn, reduce the effectiveness of the system being used as a real time system. This paper introduces a novel parameterized technique for iris segmentation. The method is based on a number of steps starting from converting grayscale eye image to a bit plane representation, selection of the most significant bit planes followed by a parameterization of the iris location resulting in an accurate segmentation of the iris from the original image. A lossless Hexadata encoding method is then applied to the data, which is based on reducing each set of six data items to a single encoded value. The tested results achieved acceptable saving bytes performance for the 21 iris square images of sizes 256x256 pixels which is about 22.4 KB on average with 0.79 sec decompression average time, with high saving bytes performance for 2 iris non-square images of sizes 640x480/2048x1536 that reached 76KB/2.2 sec, 1630 KB/4.71 sec respectively, Finally, the proposed promising techniques standard lossless JPEG2000 compression techniques with reduction about 1.2 and more in KB saving that implicitly demonstrating the power and efficiency of the suggested lossless biometric techniques.
Carbon monoxide (CO) plays an important indirect greenhouse gases due to its influences on the budgets of hydroxyl radicals (OH) and Ozone (O3). The atmospheric carbon monoxide (CO) observations can only be made on global and continental scales by remote sensing instruments situated in space. One of instrument is the Measurements of Pollution in the Troposphere (MOPITT), which is designed to measure troposphere CO and CH4 by use of a nadir-viewing geometry and was launched aboard the Earth Observing System (EOS) Terra spacecraft on 18 December 1999. Results from the analysis of the retrieved monthly (1ºх1º) spatial grid resolution, from the MOPITT data were utilized to analyze the distribution of CO surface mixing ratio in Iraq for th
... Show MoreThis paper proposes a new encryption method. It combines two cipher algorithms, i.e., DES and AES, to generate hybrid keys. This combination strengthens the proposed W-method by generating high randomized keys. Two points can represent the reliability of any encryption technique. Firstly, is the key generation; therefore, our approach merges 64 bits of DES with 64 bits of AES to produce 128 bits as a root key for all remaining keys that are 15. This complexity increases the level of the ciphering process. Moreover, it shifts the operation one bit only to the right. Secondly is the nature of the encryption process. It includes two keys and mixes one round of DES with one round of AES to reduce the performance time. The W-method deals with
... Show MoreAbstract:
Research Topic: Ruling on the sale of big data
Its objectives: a statement of what it is, importance, source and governance.
The methodology of the curriculum is inductive, comparative and critical
One of the most important results: it is not permissible to attack it and it is a valuable money, and it is permissible to sell big data as long as it does not contain data to users who are not satisfied with selling it
Recommendation: Follow-up of studies dealing with the provisions of the issue
Subject Terms
Judgment, Sale, Data, Mega, Sayings, Jurists
Referral techniques are normally employed in internet business applications. Existing frameworks prescribe things to a particular client according to client inclinations and former high evaluations. Quite a number of methods, such as cooperative filtering and content-based methodologies, dominate the architectural design of referral frameworks. Many referral schemes are domain-specific and cannot be deployed in a general-purpose setting. This study proposes a two-dimensional (User × Item)-space multimode referral scheme, having an enormous client base but few articles on offer. Additionally, the design of the referral scheme is anchored on the and articles, as expressed by a particular client, and is a combination of affi
... Show MoreThis work bases on encouraging a generous and conceivable estimation for modified an algorithm for vehicle travel times on a highway from the eliminated traffic information using set aside camera image groupings. The strategy for the assessment of vehicle travel times relies upon the distinctive verification of traffic state. The particular vehicle velocities are gotten from acknowledged vehicle positions in two persistent images by working out the distance covered all through elapsed past time doing mollification between the removed traffic flow data and cultivating a plan to unequivocally predict vehicle travel times. Erbil road data base is used to recognize road locales around road segments which are projected into the commended camera
... Show MoreThe science of information security has become a concern of many researchers, whose efforts are trying to come up with solutions and technologies that ensure the transfer of information in a more secure manner through the network, especially the Internet, without any penetration of that information, given the risk of digital data being sent between the two parties through an insecure channel. This paper includes two data protection techniques. The first technique is cryptography by using Menezes Vanstone elliptic curve ciphering system, which depends on public key technologies. Then, the encoded data is randomly included in the frame, depending on the seed used. The experimental results, using a PSNR within avera
... Show MoreIn this study, we present different methods of estimating fuzzy reliability of a two-parameter Rayleigh distribution via the maximum likelihood estimator, median first-order statistics estimator, quartile estimator, L-moment estimator, and mixed Thompson-type estimator. The mean-square error MSE as a measurement for comparing the considered methods using simulation through different values for the parameters and unalike sample sizes is used. The results of simulation show that the fuzziness values are better than the real values for all sample sizes, as well as the fuzzy reliability at the estimation of the Maximum likelihood Method, and Mixed Thompson Method perform better than the other methods in the sense of MSE, so that
... Show More<span>Distributed denial-of-service (DDoS) attack is bluster to network security that purpose at exhausted the networks with malicious traffic. Although several techniques have been designed for DDoS attack detection, intrusion detection system (IDS) It has a great role in protecting the network system and has the ability to collect and analyze data from various network sources to discover any unauthorized access. The goal of IDS is to detect malicious traffic and defend the system against any fraudulent activity or illegal traffic. Therefore, IDS monitors outgoing and incoming network traffic. This paper contains a based intrusion detection system for DDoS attack, and has the ability to detect the attack intelligently, dynami
... Show MoreThe deployment of UAVs is one of the key challenges in UAV-based communications while using UAVs for IoT applications. In this article, a new scheme for energy efficient data collection with a deadline time for the Internet of things (IoT) using the Unmanned Aerial Vehicles (UAV) is presented. We provided a new data collection method, which was set to collect IoT node data by providing an efficient deployment and mobility of multiple UAV, used to collect data from ground internet of things devices in a given deadline time. In the proposed method, data collection was done with minimum energy consumption of IoTs as well as UAVs. In order to find an optimal solution to this problem, we will first provide a mixed integer linear programming m
... Show More