The main aim of image compression is to reduce the its size to be able for transforming and storage, therefore many methods appeared to compress the image, one of these methods is "Multilayer Perceptron ". Multilayer Perceptron (MLP) method which is artificial neural network based on the Back-Propagation algorithm for compressing the image. In case this algorithm depends upon the number of neurons in the hidden layer only the above mentioned will not be quite enough to reach the desired results, then we have to take into consideration the standards which the compression process depend on to get the best results. We have trained a group of TIFF images with the size of (256*256) in our research, compressed them by using MLP for each compression process the number of neurons in the hidden layer was changing and calculating the compression ratio, mean square error and peak signal-to-noise ratio to compare the results to get the value of original image. The findings of the research was the desired results as the compression ratio was less than five and a few mean square error thus a large value of peak signal-to-noise ratio had been recorded.
Error control schemes became a necessity in network-on-chip (NoC) to improve reliability as the on-chip interconnect errors increase with the continuous shrinking of geometry. Accordingly, many researchers are trying to present multi-bit error correction coding schemes that perform a high error correction capability with the simplest design possible to minimize area and power consumption. A recent work, Multi-bit Error Correcting Coding with Reduced Link Bandwidth (MECCRLB), showed a huge reduction in area and power consumption compared to a well-known scheme, namely, Hamming product code (HPC) with Type-II HARQ. Moreover, the authors showed that the proposed scheme can correct 11 random errors which is considered a high
... Show MoreIn this research, the mechanism of cracks propagation for epoxy/ chopped carbon fibers composites have been investigated .Carbon fibers (5%, 10%, 15%, and 20%) by weight were used to reinforce epoxy resin. Bending test was carried out to evaluate the flexural strength in order to explain the mechanism of cracks propagation. It was found that, the flexural strength will increase with increasing the percentage weight for carbon fibers. At low stresses, the cracks will state at the lower surface for the specimen. Increasing the stresses will accelerate the speed of cracks until fracture accorded .The path of cracks is changed according to the distributions of carbon fibers
An automatic text summarization system mimics how humans summarize by picking the most significant sentences in a source text. However, the complexities of the Arabic language have become challenging to obtain information quickly and effectively. The main disadvantage of the traditional approaches is that they are strictly constrained (especially for the Arabic language) by the accuracy of sentence feature functions, weighting schemes, and similarity calculations. On the other hand, the meta-heuristic search approaches have a feature tha
... Show MoreIn this study, He's parallel numerical algorithm by neural network is applied to type of integration of fractional equations is Abel’s integral equations of the 1st and 2nd kinds. Using a Levenberge – Marquaradt training algorithm as a tool to train the network. To show the efficiency of the method, some type of Abel’s integral equations is solved as numerical examples. Numerical results show that the new method is very efficient problems with high accuracy.
Scheduling Timetables for courses in the big departments in the universities is a very hard problem and is often be solved by many previous works although results are partially optimal. This work implements the principle of an evolutionary algorithm by using genetic theories to solve the timetabling problem to get a random and full optimal timetable with the ability to generate a multi-solution timetable for each stage in the collage. The major idea is to generate course timetables automatically while discovering the area of constraints to get an optimal and flexible schedule with no redundancy through the change of a viable course timetable. The main contribution in this work is indicated by increasing the flexibility of generating opti
... Show MoreIn this paper, Mann-Kendall test was used to investigate the existence of possible deterministic and stochastic climatic trends in (Baghdad,Basrah,Mosul,Al-Qaim) stations. The statistical test was applied to annual monthly mean of temperatures for the period (19932009). The values of S-statistic were (62, 44, 52, 64) by comparing these values with the table of null probability values for S we get a probability of (0.002, 0.026, 0.010, 0.002) this result is less than α for the 95% confidence level (α = 0.05) indicating a significant result at this level of confidence. Concluded that an increasing trend in concentration is present at the 95% confidence level and the variance of the S-statistic is calculated and it is com
... Show MoreIn this paper, a least squares group finite element method for solving coupled Burgers' problem in 2-D is presented. A fully discrete formulation of least squares finite element method is analyzed, the backward-Euler scheme for the time variable is considered, the discretization with respect to space variable is applied as biquadratic quadrangular elements with nine nodes for each element. The continuity, ellipticity, stability condition and error estimate of least squares group finite element method are proved. The theoretical results show that the error estimate of this method is . The numerical results are compared with the exact solution and other available literature when the convection-dominated case to illustrate the effic
... Show MoreString matching is seen as one of the essential problems in computer science. A variety of computer applications provide the string matching service for their end users. The remarkable boost in the number of data that is created and kept by modern computational devices influences researchers to obtain even more powerful methods for coping with this problem. In this research, the Quick Search string matching algorithm are adopted to be implemented under the multi-core environment using OpenMP directive which can be employed to reduce the overall execution time of the program. English text, Proteins and DNA data types are utilized to examine the effect of parallelization and implementation of Quick Search string matching algorithm on multi-co
... Show MoreThe migration from IPv4 to IPv6 can not be achieved in a brief period, thus both protocols co-exist at certain years. IETF Next Generation Transition Working Group (NGtrans) developed IPv4/IPv6 transition mechanisms. Since Iraq infrastructure, including universities, companies and institutions still use IPv4 protocol only. This research article tries to highlight, discuss a required transition roadmap and extend the local knowledge and practice on IPv6. Also, it introduces a prototype model using Packet tracer (network simulator) deployed for the design and implementation of IPv6 migration. Finally, it compares and evaluates the performance of IPv6, IPv4 and dual stack using OPNET based on QoS metrics such as throughput, delay and point to
... Show More