Nanoparticles (NPs) have unique capabilities that make them an eye-opener opportunity for the upstream oil industry. Their nano-size allows them to flow within reservoir rocks without the fear of retention between micro-sized pores. Incorporating NPs with drilling and completion fluids has proved to be an effective additive that improves various properties such as mud rheology, filtration, thermal conductivity, and wellbore stability. However, the biodegradability of drilling fluid chemicals is becoming a global issue as the discharged wetted cuttings raise toxicity concerns and environmental hazards. Therefore, it is urged to utilize chemicals that tend to break down and susceptible to biodegradation. This research presents the practical application of bio-based Zinc Oxide nanoparticles (ZnO NPs) prepared chemically from celery leaf plant extract as green additive in water-based mud drilling fluid (WBM). The study aimed to evaluate the filtration and thermal stability of WBM using green-synthesized ZnO NPs. The results showed that the ZnO NPs have minimal effect of mud density, but significant improvement in mud thermal stability and filtration properties were attained with concentrations lower than 1g. The fluid loss rate was reduced by 33% with 0.45g of ZnO nanoparticles, and the thinnest mud cake was obtained as well. In terms of thermal stability, the bio-based ZnO NPs greatly enhanced the rheological properties of WBM at elevated temperatures. The rate of increment in plastic viscosity (PV) or decrement in yield point (YP) and gel strength occurred in a controllable manner compared to the rheological properties of base mud at high temperatures reaching 90°C. This study provides insight into the effect of green-synthesized ZnO nanoparticles on the performance of water-based mud and highlights their potential as an effective and environmentally friendly additive for the oil and gas industry.
Advances in digital technology and the World Wide Web has led to the increase of digital documents that are used for various purposes such as publishing and digital library. This phenomenon raises awareness for the requirement of effective techniques that can help during the search and retrieval of text. One of the most needed tasks is clustering, which categorizes documents automatically into meaningful groups. Clustering is an important task in data mining and machine learning. The accuracy of clustering depends tightly on the selection of the text representation method. Traditional methods of text representation model documents as bags of words using term-frequency index document frequency (TFIDF). This method ignores the relationship an
... Show MoreThe agent-based modeling is currently utilized extensively to analyze complex systems. It supported such growth, because it was able to convey distinct levels of interaction in a complex detailed environment. Meanwhile, agent-based models incline to be progressively complex. Thus, powerful modeling and simulation techniques are needed to address this rise in complexity. In recent years, a number of platforms for developing agent-based models have been developed. Actually, in most of the agents, often discrete representation of the environment, and one level of interaction are presented, where two or three are regarded hardly in various agent-based models. The key issue is that modellers work in these areas is not assisted by simulation plat
... Show MoreMerging biometrics with cryptography has become more familiar and a great scientific field was born for researchers. Biometrics adds distinctive property to the security systems, due biometrics is unique and individual features for every person. In this study, a new method is presented for ciphering data based on fingerprint features. This research is done by addressing plaintext message based on positions of extracted minutiae from fingerprint into a generated random text file regardless the size of data. The proposed method can be explained in three scenarios. In the first scenario the message was used inside random text directly at positions of minutiae in the second scenario the message was encrypted with a choosen word before ciphering
... Show MoreVol. 6, Issue 1 (2025)
For businesses that provide delivery services, the efficiency of the delivery process in terms of punctuality is very important. In addition to increasing customer trust, efficient route management, and selection are required to reduce vehicle fuel costs and expedite delivery. Some small and medium businesses still use conventional methods to manage delivery routes. Decisions to manage delivery schedules and routes do not use any specific methods to expedite the delivery settlement process. This process is inefficient, takes a long time, increases costs and is prone to errors. Therefore, the Dijkstra algorithm has been used to improve the delivery management process. A delivery management system was developed to help managers and drivers
... Show MoreA security system can be defined as a method of providing a form of protection to any type of data. A sequential process must be performed in most of the security systems in order to achieve good protection. Authentication can be defined as a part of such sequential processes, which is utilized in order to verify the user permission to entree and utilize the system. There are several kinds of methods utilized, including knowledge, and biometric features. The electroencephalograph (EEG) signal is one of the most widely signal used in the bioinformatics field. EEG has five major wave patterns, which are Delta, Theta, Alpha, Beta and Gamma. Every wave has five features which are amplitude, wavelength, period, speed and frequency. The linear
... Show MoreDust is a frequent contributor to health risks and changes in the climate, one of the most dangerous issues facing people today. Desertification, drought, agricultural practices, and sand and dust storms from neighboring regions bring on this issue. Deep learning (DL) long short-term memory (LSTM) based regression was a proposed solution to increase the forecasting accuracy of dust and monitoring. The proposed system has two parts to detect and monitor the dust; at the first step, the LSTM and dense layers are used to build a system using to detect the dust, while at the second step, the proposed Wireless Sensor Networks (WSN) and Internet of Things (IoT) model is used as a forecasting and monitoring model. The experiment DL system
... Show MorePredicting permeability is a cornerstone of petroleum reservoir engineering, playing a vital role in optimizing hydrocarbon recovery strategies. This paper explores the application of neural networks to predict permeability in oil reservoirs, underscoring their growing importance in addressing traditional prediction challenges. Conventional techniques often struggle with the complexities of subsurface conditions, making innovative approaches essential. Neural networks, with their ability to uncover complicated patterns within large datasets, emerge as a powerful alternative. The Quanti-Elan model was used in this study to combine several well logs for mineral volumes, porosity and water saturation estimation. This model goes be
... Show More