In this paper, we investigate and characterize the effects of multi-channel and rendezvous protocols on the connectivity of dynamic spectrum access networks using percolation theory. In particular, we focus on the scenario where the secondary nodes have plenty of vacant channels to choose from a phenomenon which we define as channel abundance. To cope with the existence of multi-channel, we use two types of rendezvous protocols: naive ones which do not guarantee a common channel and advanced ones which do. We show that, with more channel abundance, even with the use of either type of rendezvous protocols, it becomes difficult for two nodes to agree on a common channel, thereby, potentially remaining invisible to each other. We model this invisibility as a Poisson thinning process and show that invisibility is even more pronounced with channel abundance. Following the disk graph model, we represent the multiple channels as parallel edges in a graph and build a multi-layered graph (MLG) in R2. In order to study the connectivity, we show how percolation occurs in the MLG by coupling it with a typical discrete percolation. Using a Boolean model and the MLG, we study both cases of primaries' absence and presence. For both cases, we define and characterize connectivity of the secondary network in terms of the available number of channels, deployment densities, number of simultaneous transmissions per node, and communication range. When primary users are absent, we derive the critical number of channels which maintains supercriticality of the secondary network. When primary users are present, we characterize and analyze the connectivity for all the regions: channel abundance, optimal, and channel deprivation. For each region we show the requirement and the outcome of using either type of rendezvous techniques. Moreover, we find the tradeoff between deployment-density versus rendezvous probability which results in a connected network. Our results can be used to decide on the goodness of any channel rendezvous algorithm by computing the expected resultant connectivity. They also provide a guideline for achieving connectivity using minimal resources.
The purpose of this research is to investigate the impact of corrosive environment (corrosive ferric chloride of 1, 2, 5, 6% wt. at room temperature), immersion period of (48, 72, 96, 120, 144 hours), and surface roughness on pitting corrosion characteristics and use the data to build an artificial neural network and test its ability to predict the depth and intensity of pitting corrosion in a variety of conditions. Pit density and depth were calculated using a pitting corrosion test on carbon steel (C-4130). Pitting corrosion experimental tests were used to develop artificial neural network (ANN) models for predicting pitting corrosion characteristics. It was found that artificial neural network models were shown to be
... Show MorePermeability determination in Carbonate reservoir is a complex problem, due to their capability to be tight and heterogeneous, also core samples are usually only available for few wells therefore predicting permeability with low cost and reliable accuracy is an important issue, for this reason permeability predictive models become very desirable.
This paper will try to develop the permeability predictive model for one of Iraqi carbonate reservoir from core and well log data using the principle of Hydraulic Flow Units (HFUs). HFU is a function of Flow Zone Indicator (FZI) which is a good parameter to determine (HFUs).
Histogram analysis, probability analysis and Log-Log plot of Reservoir Qua
... Show More<p>The demand for internet applications has increased rapidly. Providing quality of service (QoS) requirements for varied internet application is a challenging task. One important factor that is significantly affected on the QoS service is the transport layer. The transport layer provides end-to-end data transmission across a network. Currently, the most common transport protocols used by internet application are TCP (Transmission Control Protocol) and UDP (User Datagram Protocol). Also, there are recent transport protocols such as DCCP (data congestion control protocol), SCTP (stream congestion transmission protocol), and TFRC (TCP-friendly rate control), which are in the standardization process of Internet Engineering Task
... Show More<span>As a result of numerous applications and low installation costs, wireless sensor networks (WSNs) have expanded excessively. The main concern in the WSN environment is to lower energy consumption amidst nodes while preserving an acceptable level of service quality. Using multi-mobile sinks to reduce the nodes' energy consumption have been considered as an efficient strategy. In such networks, the dynamic network topology created by the sinks mobility makes it a challenging task to deliver the data to the sinks. Thus, in order to provide efficient data dissemination, the sensor nodes will have to readjust the routes to the current position of the mobile sinks. The route re-adjustment process could result in a significant m
... Show MoreLocalization is an essential demand in wireless sensor networks (WSNs). It relies on several types of measurements. This paper focuses on positioning in 3-D space using time-of-arrival- (TOA-) based distance measurements between the target node and a number of anchor nodes. Central localization is assumed and either RF, acoustic or UWB signals are used for distance measurements. This problem is treated by using iterative gradient descent (GD), and an iterative GD-based algorithm for localization of moving sensors in a WSN has been proposed. To localize a node in 3-D space, at least four anchors are needed. In this work, however, five anchors are used to get better accuracy. In GD localization of a moving sensor, the algo
... Show MoreComputer systems and networks are being used in almost every aspect of our daily life, the security threats to computers and networks have increased significantly. Usually, password-based user authentication is used to authenticate the legitimate user. However, this method has many gaps such as password sharing, brute force attack, dictionary attack and guessing. Keystroke dynamics is one of the famous and inexpensive behavioral biometric technologies, which authenticate a user based on the analysis of his/her typing rhythm. In this way, intrusion becomes more difficult because the password as well as the typing speed must match with the correct keystroke patterns. This thesis considers static keystroke dynamics as a transparent layer of t
... Show MoreThe employment of cognitive radio (CR) is critical to the successful development of wireless communications. In this field, especially when using the multiple input multiple output (MIMO) antenna technology, energy consumption is critical. If the principal user (PU) is present, developers can utilize the energy detecting approach to tell. The researchers employed two distinct phases to conduct their research: the intense and accurate sensing stages. After the furious sensing step was completed, the PU user was identified as having a maximum or minimal energy channel. There are two situations in which the proposed algorithm's performance is tested: channels for fading AWGN and Rayleigh. When the proposed methods' simulation results a
... Show MoreOver the past few decades, the surveying fieldworks were usually carried out based on classical positioning methods for establishing horizontal and vertical geodetic networks. However, these conventional positioning techniques have many drawbacks such as time-consuming, too costly, and require massive effort. Thus, the Global Navigation Satellite System (GNSS) has been invented to fulfill the quickness, increase the accuracy, and overcome all the difficulties inherent in almost every surveying fieldwork. This research assesses the accuracy of local geodetic networks using different Global Navigation Satellite System (GNSS) techniques, such as Static, Precise Point Positioning, Post Processing Kinematic, Session method, a
... Show More