<p>The demand for internet applications has increased rapidly. Providing quality of service (QoS) requirements for varied internet application is a challenging task. One important factor that is significantly affected on the QoS service is the transport layer. The transport layer provides end-to-end data transmission across a network. Currently, the most common transport protocols used by internet application are TCP (Transmission Control Protocol) and UDP (User Datagram Protocol). Also, there are recent transport protocols such as DCCP (data congestion control protocol), SCTP (stream congestion transmission protocol), and TFRC (TCP-friendly rate control), which are in the standardization process of Internet Engineering Task
... Show MoreThis study aimed to investigate the incorporation of recycled waste compact discs (WCDs) powder in concrete mixes to replace the fine aggregate by 5%, 10%, 15% and 20%. Compared to the reference concrete mix, results revealed that using WCDs powder in concrete mixes improved the workability and the dry density. The results demonstrated that the compressive, flexural, and split tensile strengths values for the WCDs-modified concrete mixes showed tendency to increase above the reference mix. However, at 28 days curing age, the strengths values for WCDs-modified concrete mixes were comparable to those for the reference mix. The leaching test revealed that none of the WCDs constituents was detected in the leachant after 180 days. The
... Show MoreThe traditional centralized network management approach presents severe efficiency and scalability limitations in large scale networks. The process of data collection and analysis typically involves huge transfers of management data to the manager which cause considerable network throughput and bottlenecks at the manager side. All these problems processed using the Agent technology as a solution to distribute the management functionality over the network elements. The proposed system consists of the server agent that is working together with clients agents to monitor the logging (off, on) of the clients computers and which user is working on it. file system watcher mechanism is used to indicate any change in files. The results were presente
... Show MoreThe investigation of signature validation is crucial to the field of personal authenticity. The biometrics-based system has been developed to support some information security features.Aperson’s signature, an essential biometric trait of a human being, can be used to verify their identification. In this study, a mechanism for automatically verifying signatures has been suggested. The offline properties of handwritten signatures are highlighted in this study which aims to verify the authenticity of handwritten signatures whether they are real or forged using computer-based machine learning techniques. The main goal of developing such systems is to verify people through the validity of their signatures. In this research, images of a group o
... Show MoreThe main aim of image compression is to reduce the its size to be able for transforming and storage, therefore many methods appeared to compress the image, one of these methods is "Multilayer Perceptron ". Multilayer Perceptron (MLP) method which is artificial neural network based on the Back-Propagation algorithm for compressing the image. In case this algorithm depends upon the number of neurons in the hidden layer only the above mentioned will not be quite enough to reach the desired results, then we have to take into consideration the standards which the compression process depend on to get the best results. We have trained a group of TIFF images with the size of (256*256) in our research, compressed them by using MLP for each
... Show MoreImage Fusion Using A Convolutional Neural Network