Preferred Language
Articles
/
ZRdnMI8BVTCNdQwCBV8p
Iris Data Compression Based on Hexa-Data Coding
...Show More Authors

Iris research is focused on developing techniques for identifying and locating relevant biometric features, accurate segmentation and efficient computation while lending themselves to compression methods. Most iris segmentation methods are based on complex modelling of traits and characteristics which, in turn, reduce the effectiveness of the system being used as a real time system. This paper introduces a novel parameterized technique for iris segmentation. The method is based on a number of steps starting from converting grayscale eye image to a bit plane representation, selection of the most significant bit planes followed by a parameterization of the iris location resulting in an accurate segmentation of the iris from the original image. A lossless Hexadata encoding method is then applied to the data, which is based on reducing each set of six data items to a single encoded value. The tested results achieved acceptable saving bytes performance for the 21 iris square images of sizes 256x256 pixels which is about 22.4 KB on average with 0.79 sec decompression  average time, with high saving bytes performance for 2 iris non-square images of sizes 640x480/2048x1536 that reached 76KB/2.2 sec, 1630 KB/4.71 sec respectively, Finally, the proposed promising techniques standard lossless JPEG2000 compression techniques with reduction about 1.2 and more in KB saving that implicitly demonstrating the power and efficiency of the suggested lossless biometric techniques.

Crossref
View Publication
Publication Date
Wed Mar 10 2021
Journal Name
Baghdad Science Journal
Merge Operation Effect On Image Compression Using Fractal Technique
...Show More Authors

Fractal image compression gives some desirable properties like fast decoding image, and very good rate-distortion curves, but suffers from a high encoding time. In fractal image compression a partitioning of the image into ranges is required. In this work, we introduced good partitioning process by means of merge approach, since some ranges are connected to the others. This paper presents a method to reduce the encoding time of this technique by reducing the number of range blocks based on the computing the statistical measures between them . Experimental results on standard images show that the proposed method yields minimize (decrease) the encoding time and remain the quality results passable visually.

View Publication Preview PDF
Publication Date
Fri Mar 01 2019
Journal Name
Neurocomputing
A survey on video compression fast block matching algorithms
...Show More Authors

View Publication
Scopus (13)
Crossref (12)
Scopus Clarivate Crossref
Publication Date
Sun Nov 26 2017
Journal Name
Journal Of Engineering
Compression Index and Compression Ratio Prediction by Artificial Neural Networks
...Show More Authors

Information about soil consolidation is essential in geotechnical design. Because of the time and expense involved in performing consolidation tests, equations are required to estimate compression index from soil index properties. Although many empirical equations concerning soil properties have been proposed, such equations may not be appropriate for local situations. The aim of this study is to investigate the consolidation and physical properties of the cohesive soil. Artificial Neural Network (ANN) has been adapted in this investigation to predict the compression index and compression ratio using basic index properties. One hundred and ninety five consolidation results for soils tested at different construction sites

... Show More
View Publication Preview PDF
Publication Date
Mon Dec 31 2018
Journal Name
Journal Of Theoretical And Applied Information Technology (jatit)
Factors and Model for Sensitive Data Management and Protection in Information Systems’ Decision of Cloud Environment
...Show More Authors

Journal of Theoretical and Applied Information Technology is a peer-reviewed electronic research papers & review papers journal with aim of promoting and publishing original high quality research dealing with theoretical and scientific aspects in all disciplines of IT (Informaiton Technology

View Publication
Publication Date
Fri Oct 01 2010
Journal Name
2010 Ieee Symposium On Industrial Electronics And Applications (isiea)
Distributed t-way test suite data generation using exhaustive search method with map and reduce framework
...Show More Authors

View Publication
Scopus (2)
Crossref (2)
Scopus Crossref
Publication Date
Sat Jan 01 2011
Journal Name
International Journal Of Data Analysis Techniques And Strategies
A class of efficient and modified testimators for the mean of normal distribution using complete data
...Show More Authors

View Publication
Scopus (9)
Crossref (2)
Scopus Crossref
Publication Date
Sat Feb 01 2020
Journal Name
Iop Conference Series: Materials Science And Engineering
Revealing the potentials of 3D modelling techniques; a comparison study towards data fusion from hybrid sensors
...Show More Authors
Abstract<p>The vast advantages of 3D modelling industry have urged competitors to improve capturing techniques and processing pipelines towards minimizing labour requirements, saving time and reducing project risk. When it comes to digital 3D documentary and conserving projects, laser scanning and photogrammetry are compared to choose between the two. Since both techniques have pros and cons, this paper approaches the potential issues of individual techniques in terms of time, budget, accuracy, density, methodology and ease to use. Terrestrial laser scanner and close-range photogrammetry are tested to document a unique invaluable artefact (Lady of Hatra) located in Iraq for future data fusion sc</p> ... Show More
View Publication
Scopus (8)
Crossref (4)
Scopus Clarivate Crossref
Publication Date
Thu Jan 01 2015
Journal Name
Journal Of Engineering
A Visual Interface Design for Evaluating the Quality of Google Map Data for some Engineering Applications
...Show More Authors

Today, there are large amounts of geospatial data available on the web such as Google Map (GM), OpenStreetMap (OSM), Flickr service, Wikimapia and others. All of these services called open source geospatial data. Geospatial data from different sources often has variable accuracy due to different data collection methods; therefore data accuracy may not meet the user requirement in varying organization. This paper aims to develop a tool to assess the quality of GM data by comparing it with formal data such as spatial data from Mayoralty of Baghdad (MB). This tool developed by Visual Basic language, and validated on two different study areas in Baghdad / Iraq (Al-Karada and Al- Kadhumiyah). The positional accuracy was asses

... Show More
View Publication
Publication Date
Sun Mar 04 2012
Journal Name
Baghdad Science Journal
Using fuzzy logic for estimating monthly pan evaporation from meteorological data in Emara/ South of Iraq
...Show More Authors

Evaporation is one of the major components of the hydrological cycle in the nature, thus its accurate estimation is so important in the planning and management of the irrigation practices and to assess water availability and requirements. The aim of this study is to investigate the ability of fuzzy inference system for estimating monthly pan evaporation form meteorological data. The study has been carried out depending on 261 monthly measurements of each of temperature (T), relative humidity (RH), and wind speed (W) which have been available in Emara meteorological station, southern Iraq. Three different fuzzy models comprising various combinations of monthly climatic variables (temperature, wind speed, and relative humidity) were developed

... Show More
View Publication Preview PDF
Crossref
Publication Date
Wed Apr 25 2018
Journal Name
Ibn Al-haitham Journal For Pure And Applied Sciences
Using Approximation Non-Bayesian Computation with Fuzzy Data to Estimation Inverse Weibull Parameters and Reliability Function
...Show More Authors

        In real situations all observations and measurements are not exact numbers but more or less non-exact, also called fuzzy. So, in this paper, we use approximate non-Bayesian computational methods to estimate inverse Weibull parameters and reliability function with fuzzy data. The maximum likelihood and moment estimations are obtained as non-Bayesian estimation. The maximum likelihood estimators have been derived numerically based on two iterative techniques namely “Newton-Raphson” and the “Expectation-Maximization” techniques. In addition, we provide compared numerically through Monte-Carlo simulation study to obtained estimates of the parameters and reliability function i

... Show More
View Publication Preview PDF
Crossref