A robust video-bitrate adaptive scheme at client-aspect plays a significant role in keeping a good quality of video streaming technology experience. Video quality affects the amount of time the video has turned off playing due to the unfilled buffer state. Therefore to maintain a video streaming continuously with smooth bandwidth fluctuation, a video buffer structure based on adapting the video bitrate is considered in this work. Initially, the video buffer structure is formulated as an optimal control-theoretic problem that combines both video bitrate and video buffer feedback signals. While protecting the video buffer occupancy from exceeding the limited operating level can provide continuous video streaming, it may also cause a video bitrate oscillation. So the video buffer structure is adjusted by adding two thresholds as operating points for overflow and underflow states to filter the impact of throughput fluctuation on video buffer occupancy level. Then a bandwidth prediction algorithm is proposed for enhancing the performance of video bitrate adaptation. This algorithm's work depends on the current video buffer level, video bitrate of the previous segment, and iterative throughput measurements to predict the best video bitrate for the next segment. Simulation results show that reserving a bandwidth margin is better in adapting the video bitrate under bandwidth variation and then reducing the risk of video playback freezing. Simulation results proved that the playback freezing happens two times: firstly, when there is no bandwidth margin used and secondly, when the bandwidth margin is high while smooth video bitrate is obtained with moderate value. The proposed scheme is compared with other two schemes such as smoothed throughput rate (STR) and Buffer Based Rate (BBR) in terms of prediction error, QoE preferences, buffer size, and startup delay time, then the proposed scheme outperforms these schemes in attaining smooth video bitrates and continuous video playback.
The current research's problem includes the impact of cognitive reappraisal and reformulate on self-experience of emotional response and its negative feelings and the activity of cognitive reappraisal in changing response. The aim of this research is to detect the relation between adaptive response and cognitive reappraisal upon students of secondary school, and to find differences in gender and stage. The sample contained male and female student for the year(2022-2023) and consists of (480) students (240) male and (240) female in the karkh education/ 1 To achieve this aims researcher used descriptive method and to measure the two variables researcher built a scale for adaptive response according to theory of compound emotion (Barrett,20
... Show MoreHeuristic approaches are traditionally applied to find the optimal size and optimal location of Flexible AC Transmission Systems (FACTS) devices in power systems. Genetic Algorithm (GA) technique has been applied to solve power engineering optimization problems giving better results than classical methods. This paper shows the application of GA for optimal sizing and allocation of a Static Compensator (STATCOM) in a power system. STATCOM devices used to increase transmission systems capacity and enhance voltage stability by regulate the voltages at its terminal by controlling the amount of reactive power injected into or absorbed from the power system. IEEE 5-bus standard system is used as an example to illustrate the te
... Show MoreThe aim of this study is to develop a novel framework for managing risks in smart supply chains by enhancing business continuity and resilience against potential disruptions. This research addresses the growing uncertainty in supply chain environments, driven by both natural phenomena-such as pandemics and earthquakes—and human-induced events, including wars, political upheavals, and societal transformations. Recognizing that traditional risk management approaches are insufficient in such dynamic contexts, the study proposes an adaptive framework that integrates proactive and remedial measures for effective risk mitigation. A fuzzy risk matrix is employed to assess and analyze uncertainties, facilitating the identification of disr
... Show MoreThis abstract focuses on the significance of wireless body area networks (WBANs) as a cutting-edge and self-governing technology, which has garnered substantial attention from researchers. The central challenge faced by WBANs revolves around upholding quality of service (QoS) within rapidly evolving sectors like healthcare. The intricate task of managing diverse traffic types with limited resources further compounds this challenge. Particularly in medical WBANs, the prioritization of vital data is crucial to ensure prompt delivery of critical information. Given the stringent requirements of these systems, any data loss or delays are untenable, necessitating the implementation of intelligent algorithms. These algorithms play a pivota
... Show MoreEvolutionary algorithms are better than heuristic algorithms at finding protein complexes in protein-protein interaction networks (PPINs). Many of these algorithms depend on their standard frameworks, which are based on topology. Further, many of these algorithms have been exclusively examined on networks with only reliable interaction data. The main objective of this paper is to extend the design of the canonical and topological-based evolutionary algorithms suggested in the literature to cope with noisy PPINs. The design of the evolutionary algorithm is extended based on the functional domain of the proteins rather than on the topological domain of the PPIN. The gene ontology annotation in each molecular function, biological proce
... Show MoreBootstrap is one of an important re-sampling technique which has given the attention of researches recently. The presence of outliers in the original data set may cause serious problem to the classical bootstrap when the percentage of outliers are higher than the original one. Many methods are proposed to overcome this problem such Dynamic Robust Bootstrap for LTS (DRBLTS) and Weighted Bootstrap with Probability (WBP). This paper try to show the accuracy of parameters estimation by comparison the results of both methods. The bias , MSE and RMSE are considered. The criterion of the accuracy is based on the RMSE value since the method that provide us RMSE value smaller than other is con
... Show MoreAbstract
The study aimed to prepare a practical guide for procedures for auditing the strategies of municipal institutions in achieving sustainable development by adopting the idea of the audit matrix through which a classified report is prepared according to the dimensions of sustainable development, by preparing a specialized audit program for the purpose of auditing strategies for achieving sustainable development and emptying the results of the application of each of the paragraphs The program in the audit matrix that was prepared for the purpose of determining the impact of each observation and linkin
... Show MoreIncremental Sheet Metal Forming (ISMF) is a modern sheet metal forming technology which offers the possibility of manufacturing 3D complex parts of thin sheet metals using the CNC milling machine. The surface quality is a very important aspect in any manufacturing process. Therefore, this study focuses on the resultant residual stresses by forming parameters, namely; (tool shape, step over, feed rate, and slope angle) using Taguchi method for the products formed by single point incremental forming process (SPIF). For evaluating the surface quality, practical experiments to produce pyramid like shape have been implemented on aluminum sheets (AA1050) for thickness (0.9) mm. Three types of tool shape used in this work, the spherical tool ga
... Show MoreDeep learning (DL) plays a significant role in several tasks, especially classification and prediction. Classification tasks can be efficiently achieved via convolutional neural networks (CNN) with a huge dataset, while recurrent neural networks (RNN) can perform prediction tasks due to their ability to remember time series data. In this paper, three models have been proposed to certify the evaluation track for classification and prediction tasks associated with four datasets (two for each task). These models are CNN and RNN, which include two models (Long Short Term Memory (LSTM)) and GRU (Gated Recurrent Unit). Each model is employed to work consequently over the two mentioned tasks to draw a road map of deep learning mod
... Show MoreThe permeability is the most important parameter that indicates how efficient the reservoir fluids flow through the rock pores to the wellbore. Well-log evaluation and core measurements techniques are typically used to estimate it. In this paper, the permeability has been predicted by using classical and Flow zone indicator methods. A comparison between the two methods shows the superiority of the FZI method correlations, these correlations can be used to estimate permeability in un-cored wells with a good approximation.