Cryptography is the process of transforming message to avoid an unauthorized access of data. One of the main problems and an important part in cryptography with secret key algorithms is key. For higher level of secure communication key plays an important role. For increasing the level of security in any communication, both parties must have a copy of the secret key which, unfortunately, is not that easy to achieve. Triple Data Encryption Standard algorithm is weak due to its weak key generation, so that key must be reconfigured to make this algorithm more secure, effective, and strong. Encryption key enhances the Triple Data Encryption Standard algorithm securities. This paper proposed a combination of two efficient encryption algorithms to satisfy the purpose of information security by adding a new level of security to Triple Data Encryption Standard algorithm using Nth Degree Truncated Polynomial Ring Unit algorithm. This aim achieved by adding two new key functions, the first one is Enckey(), and the second one is Deckey() for encryption and decryption key of Triple Data Encryption Standard to make this algorithm more stronger. The obtained results of this paper also have good resistance against brute-force attack which makes the system more effective by applying Nth Degree Truncated Polynomial Ring Unit algorithm to encrypt and decrypt key of Triple Data Encryption Standard. Also, these modifications enhance the degree of complexity, increase key search space, and make the ciphered message difficult to be cracked by the attacker.
The current research discussed biophysics data as a theoretical and applied knowledge base linking industrial design with the natural sciences at the level of applied strategies through which we can enrich the knowledge base of industrial design. The research focused on two main aspects of the scientific references for biophysics, namely: electromagnetism, and biomechanics. According to the performance and functional applications in designing the functions of industrial products at the electromagnetic level, it was found that remote sensing applications: such as fire sensors that were adopted from the insect (Black Beetle) and that their metaphors enable them to hear fire, and collision sensors, which were adopted from the insect
... Show MoreMixed-effects conditional logistic regression is evidently more effective in the study of qualitative differences in longitudinal pollution data as well as their implications on heterogeneous subgroups. This study seeks that conditional logistic regression is a robust evaluation method for environmental studies, thru the analysis of environment pollution as a function of oil production and environmental factors. Consequently, it has been established theoretically that the primary objective of model selection in this research is to identify the candidate model that is optimal for the conditional design. The candidate model should achieve generalizability, goodness-of-fit, parsimony and establish equilibrium between bias and variab
... Show MoreThe purpose of this research is defining the main factors influencing on decision of management system on sensitive data in cloud. The framework is proposed to enhance management information systems decision on sensitive information in cloud environment. The structured interview with several security experts working on cloud computing security to investigate the main objective of framework and suitability of instrument, a pilot study conducts to test the instrument. The validity and reliability test results expose that study can be expanded and lead to final framework validation. This framework using multilevel related to Authorization, Authentication, Classification and identity anonymity, and save and verify, to enhance management
... Show MoreSpatial data analysis is performed in order to remove the skewness, a measure of the asymmetry of the probablitiy distribution. It also improve the normality, a key concept of statistics from the concept of normal distribution “bell shape”, of the properties like improving the normality porosity, permeability and saturation which can be are visualized by using histograms. Three steps of spatial analysis are involved here; exploratory data analysis, variogram analysis and finally distributing the properties by using geostatistical algorithms for the properties. Mishrif Formation (unit MB1) in Nasiriya Oil Field was chosen to analyze and model the data for the first eight wells. The field is an anticline structure with northwest- south
... Show MoreThis research dealt with study of cladistics taxonomy of five species related to the genus Rumex L. and Polygonum L. from family polygonaceae in Iraq by using Mesquite software V.2.75. This research support strongly delimiting the species P. aviculare L. and P. lapathifolia L.as suggested in floras publication while R. dentatus L. is setted in single group whereas R. vesicarius L. and R. conglomeratus Murray were included in the same group. Also, this study involved characteristics of shape, dimensions, color, and ornamentation of seeds and fruits as the seed forms were ranging from lenticular to trigonous. In terms of size calculations, the seeds of R. vesicarius was recorded the higher range (4.0- 4.5) mm in length w
... Show MorePurpose: to demonstrate the possibility of moving to electronic data exchange dimensions (regulatory requirements, technical requirements, human requirements, senior management support) to simplify the work procedures dimensions (modern procedures, clarity of procedures, short procedures, availability of information and means required. The simplicity of the models used because of its importance to keep abreast of recent developments in the service of municipal works through the application of electronic data interchange, which simplifies procedures and out of the routine in the performance of the work of municipal departments has developed. It was applied to Municipality (Hilla) so that the transformation
... Show MoreIn this study, we present different methods of estimating fuzzy reliability of a two-parameter Rayleigh distribution via the maximum likelihood estimator, median first-order statistics estimator, quartile estimator, L-moment estimator, and mixed Thompson-type estimator. The mean-square error MSE as a measurement for comparing the considered methods using simulation through different values for the parameters and unalike sample sizes is used. The results of simulation show that the fuzziness values are better than the real values for all sample sizes, as well as the fuzzy reliability at the estimation of the Maximum likelihood Method, and Mixed Thompson Method perform better than the other methods in the sense of MSE, so that
... Show MoreRealizing the full potential of wireless sensor networks (WSNs) highlights many design issues, particularly the trade-offs concerning multiple conflicting improvements such as maximizing the route overlapping for efficient data aggregation and minimizing the total link cost. While the issues of data aggregation routing protocols and link cost function in a WSNs have been comprehensively considered in the literature, a trade-off improvement between these two has not yet been addressed. In this paper, a comprehensive weight for trade-off between different objectives has been employed, the so-called weighted data aggregation routing strategy (WDARS) which aims to maximize the overlap routes for efficient data aggregation and link cost
... Show MoreIn the current study, 2D seismic data in west An-Najaf (WN-36 line) were received after many steps of processing by Oil Exploration Company in 2018. Surface Consistent Amplitude Compensation (SCAC) was applied on the seismic data. The processing sequence in our study started by sorting data in a common mid-point (CMP) gather, in order to apply the velocity analysis using Interactive Velocity Analysis Application (INVA) with Omega system. Semblance of velocity was prepared to preform normal move-out (NMO) vs. Time. Accurate root mean square velocity (VRMS) was selected, which was controlled by flatness of the primary events. The resultant seismic velocity section for the study area shows that the veloci
... Show MoreThe Arabic grammatical theory is characterized by the characteristics that distinguish it from other languages. It is based on the following equation: In its entirety a homogeneous linguistic system that blends with the social nature of the Arab, his beliefs, and his culture.
This means that this theory was born naturally, after the labor of maintaining an integrated inheritance, starting with its legal text (the Koran), and ends with its features of multiple attributes.
Saber was carrying the founding crucible of that theory, which takes over from his teacher, Hebron, to be built on what it has reached. It is redundant to point to his location and the status of his book.
So came to my research tagged: (c