The Internet is providing vital communications between millions of individuals. It is also more and more utilized as one of the commerce tools; thus, security is of high importance for securing communications and protecting vital information. Cryptography algorithms are essential in the field of security. Brute force attacks are the major Data Encryption Standard attacks. This is the main reason that warranted the need to use the improved structure of the Data Encryption Standard algorithm. This paper proposes a new, improved structure for Data Encryption Standard to make it secure and immune to attacks. The improved structure of Data Encryption Standard was accomplished using standard Data Encryption Standard with a new way of two key gene
... Show MoreJPEG is most popular image compression and encoding, this technique is widely used in many applications (images, videos and 3D animations). Meanwhile, researchers are very interested to develop this massive technique to compress images at higher compression ratios with keeping image quality as much as possible. For this reason in this paper we introduce a developed JPEG based on fast DCT and removed most of zeros and keeps their positions in a transformed block. Additionally, arithmetic coding applied rather than Huffman coding. The results showed up, the proposed developed JPEG algorithm has better image quality than traditional JPEG techniques.
Background: This in vitro study measure and compare the effect of light curing tip distance on the depth of cure by measuring vickers microhardness value on two recently launched bulk fill resin based composites Tetric EvoCeram Bulk Fill and Surefil SDR Flow with 4 mm thickness in comparison to Filtek Z250 Universal Restorative with 2 mm thickness. In addition, measure and compare the bottom to top microhardness ratio with different light curing tip distances. Materials and Method: One hundred fifty composite specimens were obtained from two cylindrical plastic molds the first one for bulk fill composites (Tetric EvoCeram Bulk Fill and Surefil SDR Flow) with 4 mm diameter and 4 mm depth, the second one for Filtek Z250 Universal Restorative
... Show MoreABSTRACT Porous silicon has been produced in this work by photochemical etching process (PC). The irradiation has been achieved using ordinary light source (150250 W) power and (875 nm) wavelength. The influence of various irradiation times and HF concentration on porosity of PSi material was investigated by depending on gravimetric measurements. The I-V and C-V characteristics for CdS/PSi structure have been investigated in this work too.
In this research, the focus was placed on estimating the parameters of the Hypoexponential distribution function using the maximum likelihood method and genetic algorithm. More than one standard, including MSE, has been adopted for comparison by Using the simulation method
The research utilizes data produced by the Local Urban Management Directorate in Najaf and the imagery data from the Landsat 9 satellite, after being processed by the GIS tool. The research follows a descriptive and analytical approach; we integrated the Markov chain analysis and the cellular automation approach to predict transformations in city structure as a result of changes in land utilization. The research also aims to identify approaches to detect post-classification transformations in order to determine changes in land utilization. To predict the future land utilization in the city of Kufa, and to evaluate data accuracy, we used the Kappa Indicator to determine the potential applicability of the probability matrix that resulted from
... Show MoreArtificial Intelligence Algorithms have been used in recent years in many scientific fields. We suggest employing artificial TABU algorithm to find the best estimate of the semi-parametric regression function with measurement errors in the explanatory variables and the dependent variable, where measurement errors appear frequently in fields such as sport, chemistry, biological sciences, medicine, and epidemiological studies, rather than an exact measurement.
The estimation of the parameters of linear regression is based on the usual Least Square method, as this method is based on the estimation of several basic assumptions. Therefore, the accuracy of estimating the parameters of the model depends on the validity of these hypotheses. The most successful technique was the robust estimation method which is minimizing maximum likelihood estimator (MM-estimator) that proved its efficiency in this purpose. However, the use of the model becomes unrealistic and one of these assumptions is the uniformity of the variance and the normal distribution of the error. These assumptions are not achievable in the case of studying a specific problem that may include complex data of more than one model. To
... Show More