Abstract Background: This study is aimed to assess the maxillary incisors’ root position, angulation, and buccal alveolar bone thickness in both genders and different classes of malocclusion using cone‑beam computed tomography (CBCT). Materials and Methods: Two hundred and six CBCT images were gathered and analyzed by three‑dimensional On‑Demand software to measure the variables of 803 maxillary central and lateral incisors. Genders and class difference was determined by unpaired t‑test, one‑way ANOVA, and Chi‑square tests. Results: Buccal root position of the maxillary incisors accounted for in the majority of the cases followed by the middle and palatal positions. The thickness of alveolar bone appears to have nearly the same pattern of decreasing in the mean values above the level of 2 mm from the crest of the bone up to the 6 mm level then increase in the apex of the root. The angle between the long axis of the maxillary incisors and the corresponding alveolar bone is higher significantly in class II followed by class I and III with no significant gender difference. Conclusions: most of the maxillary incisors examined were located in a very close position to the buccal cortical plate and covered by a thin buccal bone wall. The apparent association was noted between the incisors’ root position and angulation in the alveolar bone with the buccal
This research aims to distinguish the reef environment from the non-reef environment. The Oligocene-Miocene-succussion in western Iraq was selected as a case study, represented by the reefal limestone facies of the Anah Formation (Late Oligocene) deposited in reef-back reef environments, dolomitic limestone of the Euphrates Formation (Early Miocene) deposited in open sea environments, and gypsiferous marly limestone of the Fatha Formation (Middle Miocene) deposited in a lagoonal environment. The content of the rare earth elements (REEs) (La, Ce, Pr, Nd, Sm, Eu, Gd, Tb, Dy, Er, Ho, Tm, Yb, Lu, and Y) in reef facies appear to be much lower than of those in the non-reef facies. The open sea facies have a low content of REEs due to bein
... Show MoreThe goal of this research is to develop a numerical model that can be used to simulate the sedimentation process under two scenarios: first, the flocculation unit is on duty, and second, the flocculation unit is out of commission. The general equation of flow and sediment transport were solved using the finite difference method, then coded using Matlab software. The result of this study was: the difference in removal efficiency between the coded model and operational model for each particle size dataset was very close, with a difference value of +3.01%, indicating that the model can be used to predict the removal efficiency of a rectangular sedimentation basin. The study also revealed
<span>Digital audio is required to transmit large sizes of audio information through the most common communication systems; in turn this leads to more challenges in both storage and archieving. In this paper, an efficient audio compressive scheme is proposed, it depends on combined transform coding scheme; it is consist of i) bi-orthogonal (tab 9/7) wavelet transform to decompose the audio signal into low & multi high sub-bands, ii) then the produced sub-bands passed through DCT to de-correlate the signal, iii) the product of the combined transform stage is passed through progressive hierarchical quantization, then traditional run-length encoding (RLE), iv) and finally LZW coding to generate the output mate bitstream.
... Show MoreThe aim of this research is to compare traditional and modern methods to obtain the optimal solution using dynamic programming and intelligent algorithms to solve the problems of project management.
It shows the possible ways in which these problems can be addressed, drawing on a schedule of interrelated and sequential activities And clarifies the relationships between the activities to determine the beginning and end of each activity and determine the duration and cost of the total project and estimate the times used by each activity and determine the objectives sought by the project through planning, implementation and monitoring to maintain the budget assessed
... Show MoreIn this study, a fast block matching search algorithm based on blocks' descriptors and multilevel blocks filtering is introduced. The used descriptors are the mean and a set of centralized low order moments. Hierarchal filtering and MAE similarity measure were adopted to nominate the best similar blocks lay within the pool of neighbor blocks. As next step to blocks nomination the similarity of the mean and moments is used to classify the nominated blocks and put them in one of three sub-pools, each one represents certain nomination priority level (i.e., most, less & least level). The main reason of the introducing nomination and classification steps is a significant reduction in the number of matching instances of the pixels belong to the c
... Show MoreTo maintain the security and integrity of data, with the growth of the Internet and the increasing prevalence of transmission channels, it is necessary to strengthen security and develop several algorithms. The substitution scheme is the Playfair cipher. The traditional Playfair scheme uses a small 5*5 matrix containing only uppercase letters, making it vulnerable to hackers and cryptanalysis. In this study, a new encryption and decryption approach is proposed to enhance the resistance of the Playfair cipher. For this purpose, the development of symmetric cryptography based on shared secrets is desired. The proposed Playfair method uses a 5*5 keyword matrix for English and a 6*6 keyword matrix for Arabic to encrypt the alphabets of
... Show MoreA new algorithm is proposed to compress speech signals using wavelet transform and linear predictive coding. Signal compression based on the concept of selecting a small number of approximation coefficients after they are compressed by the wavelet decomposition (Haar and db4) at a suitable chosen level and ignored details coefficients, and then approximation coefficients are windowed by a rectangular window and fed to the linear predictor. Levinson Durbin algorithm is used to compute LP coefficients, reflection coefficients and predictor error. The compress files contain LP coefficients and previous sample. These files are very small in size compared to the size of the original signals. Compression ratio is calculated from the size of th
... Show More