International Science Index
The Origin, Diffusion and a Comparison of Ordinary Differential Equations Numerical Solutions Used by SIR Model in Order to Predict SARS-CoV-2 in Nordic Countries
SARS-CoV-2 virus is currently one of the most
infectious pathogens for humans. It started in China at the end of
2019 and now it is spread in all over the world. The origin and
diffusion of the SARS-CoV-2 epidemic, is analysed based on the
discussion of viral phylogeny theory. With the aim of understanding
the spread of infection in the affected countries, it is crucial to
modelize the spread of the virus and simulate its activity. In this
paper, the prediction of coronavirus outbreak is done by using SIR
model without vital dynamics, applying different numerical technique
solving ordinary differential equations (ODEs). We find out that ABM
and MRT methods perform better than other techniques and that the
activity of the virus will decrease in April but it never cease (for
some time the activity will remain low) and the next cycle will start
in the middle July 2020 for Norway and Denmark, and October 2020
for Sweden, and September for Finland.
Enhancing Temporal Extrapolation of Wind Speed Using a Hybrid Technique: A Case Study in West Coast of Denmark
The demand for renewable energy is significantly increasing, major investments are being supplied to the wind power generation industry as a leading source of clean energy. The wind energy sector is entirely dependable and driven by the prediction of wind speed, which by the nature of wind is very stochastic and widely random. This s0tudy employs deep multi-fidelity Gaussian process regression, used to predict wind speeds for medium term time horizons. Data of the RUNE experiment in the west coast of Denmark were provided by the Technical University of Denmark, which represent the wind speed across the study area from the period between December 2015 and March 2016. The study aims to investigate the effect of pre-processing the data by denoising the signal using empirical wavelet transform (EWT) and engaging the vector components of wind speed to increase the number of input data layers for data fusion using deep multi-fidelity Gaussian process regression (GPR). The outcomes were compared using root mean square error (RMSE) and the results demonstrated a significant increase in the accuracy of predictions which demonstrated that using vector components of the wind speed as additional predictors exhibits more accurate predictions than strategies that ignore them, reflecting the importance of the inclusion of all sub data and pre-processing signals for wind speed forecasting models.
Constructing a Bayesian Network for Solar Energy in Egypt Using Life Cycle Analysis and Machine Learning Algorithms
In an era where machines run and shape our world, the need for a stable, non-ending source of energy emerges. In this study, the focus was on the solar energy in Egypt as a renewable source, the most important factors that could affect the solar energy’s market share throughout its life cycle production were analyzed and filtered, the relationships between them were derived before structuring a Bayesian network. Also, forecasted models were built for multiple factors to predict the states in Egypt by 2035, based on historical data and patterns, to be used as the nodes’ states in the network. 37 factors were found to might have an impact on the use of solar energy and then were deducted to 12 factors that were chosen to be the most effective to the solar energy’s life cycle in Egypt, based on surveying experts and data analysis, some of the factors were found to be recurring in multiple stages. The presented Bayesian network could be used later for scenario and decision analysis of using solar energy in Egypt, as a stable renewable source for generating any type of energy needed.
Increasing the Forecasting Fidelity of Current Collection System Operating Capability by Means of Contact Pressure Simulation Modelling
Current collection quality is one of the limiting factors when increasing trains movement speed in the rail sector. With the movement speed growth, the impact forces on the current collector from the rolling stock and the aerodynamic influence increase, which leads to the spread in the contact pressure values, separation of the current collector head from the contact wire, contact arcing and excessive wear of the contact elements. The upcoming trend in resolving this issue is the use of the automatic control systems providing stabilization of the contact pressure value. The present paper considers the features of the contemporary automatic control systems of the current collector’s pressure; their major disadvantages have been stated. A scheme of current collector pressure automatic control has been proposed, distinguished by a proactive influence on undesirable effects. A mathematical model of contact strips wearing has been presented, obtained in accordance with the provisions of the central composition rotatable design program. The analysis of the obtained dependencies has been carried out. The procedures for determining the optimal current collector pressure on the contact wire and the pressure control principle in the pneumatic drive have been described.
Time Series Simulation by Conditional Generative Adversarial Net
Generative Adversarial Net (GAN) has proved to be a powerful machine learning tool in image data analysis and generation. In this paper, we propose to use Conditional Generative Adversarial Net (CGAN) to learn and simulate time series data. The conditions include both categorical and continuous variables with different auxiliary information. Our simulation studies show that CGAN has the capability to learn different types of normal and heavy-tailed distributions, as well as dependent structures of different time series. It also has the capability to generate conditional predictive distributions consistent with training data distributions. We also provide an in-depth discussion on the rationale behind GAN and the neural networks as hierarchical splines to establish a clear connection with existing statistical methods of distribution generation. In practice, CGAN has a wide range of applications in market risk and counterparty risk analysis: it can be applied to learn historical data and generate scenarios for the calculation of Value-at-Risk (VaR) and Expected Shortfall (ES), and it can also predict the movement of the market risk factors. We present a real data analysis including a backtesting to demonstrate that CGAN can outperform Historical Simulation (HS), a popular method in market risk analysis to calculate VaR. CGAN can also be applied in economic time series modeling and forecasting. In this regard, we have included an example of hypothetical shock analysis for economic models and the generation of potential CCAR scenarios by CGAN at the end of the paper.
Forecasting Stock Indexes Using Bayesian Additive Regression Tree
Forecasting the stock market is a very challenging task. Various economic indicators such as GDP, exchange rates, interest rates, and unemployment have a substantial impact on the stock market. Time series models are the traditional methods used to predict stock market changes. In this paper, a machine learning method, Bayesian Additive Regression Tree (BART) is used in predicting stock market indexes based on multiple economic indicators. BART can be used to model heterogeneous treatment effects, and thereby works well when models are misspecified. It also has the capability to handle non-linear main effects and multi-way interactions without much input from financial analysts. In this research, BART is proposed to provide a reliable prediction on day-to-day stock market activities. By comparing the analysis results from BART and with time series method, BART can perform well and has better prediction capability than the traditional methods.
The Non-Stationary BINARMA(1,1) Process with Poisson Innovations: An Application on Accident Data
This paper considers the modelling of a non-stationary
bivariate integer-valued autoregressive moving average of order
one (BINARMA(1,1)) with correlated Poisson innovations. The
BINARMA(1,1) model is specified using the binomial thinning
operator and by assuming that the cross-correlation between the
two series is induced by the innovation terms only. Based on
these assumptions, the non-stationary marginal and joint moments
of the BINARMA(1,1) are derived iteratively by using some initial
stationary moments. As regards to the estimation of parameters of
the proposed model, the conditional maximum likelihood (CML)
estimation method is derived based on thinning and convolution
properties. The forecasting equations of the BINARMA(1,1) model
are also derived. A simulation study is also proposed where
BINARMA(1,1) count data are generated using a multivariate
Poisson R code for the innovation terms. The performance of
the BINARMA(1,1) model is then assessed through a simulation
experiment and the mean estimates of the model parameters obtained
are all efficient, based on their standard errors. The proposed model
is then used to analyse a real-life accident data on the motorway in
Mauritius, based on some covariates: policemen, daily patrol, speed
cameras, traffic lights and roundabouts. The BINARMA(1,1) model
is applied on the accident data and the CML estimates clearly indicate
a significant impact of the covariates on the number of accidents on
the motorway in Mauritius. The forecasting equations also provide
reliable one-step ahead forecasts.
Forecast of Polyethylene Properties in the Gas Phase Polymerization Aided by Neural Network
A major problem that affects the quality control of polymer in the industrial polymerization is the lack of suitable on-line measurement tools to evaluate the properties of the polymer such as melt and density indices. Controlling the polymerization in ordinary method is performed manually by taking samples, measuring the quality of polymer in the lab and registry of results. This method is highly time consuming and leads to producing large number of incompatible products. An online application for estimating melt index and density proposed in this study is a neural network based on the input-output data of the polyethylene production plant. Temperature, the level of reactors' bed, the intensity of ethylene mass flow, hydrogen and butene-1, the molar concentration of ethylene, hydrogen and butene-1 are used for the process to establish the neural model. The neural network is taught based on the actual operational data and back-propagation and Levenberg-Marquart techniques. The simulated results indicate that the neural network process model established with three layers (one hidden layer) for forecasting the density and the four layers for the melt index is able to successfully predict those quality properties.
River Stage-Discharge Forecasting Based on Multiple-Gauge Strategy Using EEMD-DWT-LSSVM Approach
This study presented hybrid pre-processing approach along with a conceptual model to enhance the accuracy of river discharge prediction. In order to achieve this goal, Ensemble Empirical Mode Decomposition algorithm (EEMD), Discrete Wavelet Transform (DWT) and Mutual Information (MI) were employed as a hybrid pre-processing approach conjugated to Least Square Support Vector Machine (LSSVM). A conceptual strategy namely multi-station model was developed to forecast the Souris River discharge more accurately. The strategy used herein was capable of covering uncertainties and complexities of river discharge modeling. DWT and EEMD was coupled, and the feature selection was performed for decomposed sub-series using MI to be employed in multi-station model. In the proposed feature selection method, some useless sub-series were omitted to achieve better performance. Results approved efficiency of the proposed DWT-EEMD-MI approach to improve accuracy of multi-station modeling strategies.
Improving Flash Flood Forecasting with a Bayesian Probabilistic Approach: A Case Study on the Posina Basin in Italy
The Flash Flood Guidance (FFG) provides the rainfall amount of a given duration necessary to cause flooding. The approach is based on the development of rainfall-runoff curves, which helps us to find out the rainfall amount that would cause flooding. An alternative approach, mostly experimented with Italian Alpine catchments, is based on determining threshold discharges from past events and on finding whether or not an oncoming flood has its magnitude more than some critical discharge thresholds found beforehand. Both approaches suffer from large uncertainties in forecasting flash floods as, due to the simplistic approach followed, the same rainfall amount may or may not cause flooding. This uncertainty leads to the question whether a probabilistic model is preferable over a deterministic one in forecasting flash floods. We propose the use of a Bayesian probabilistic approach in flash flood forecasting. A prior probability of flooding is derived based on historical data. Additional information, such as antecedent moisture condition (AMC) and rainfall amount over any rainfall thresholds are used in computing the likelihood of observing these conditions given a flash flood has occurred. Finally, the posterior probability of flooding is computed using the prior probability and the likelihood. The variation of the computed posterior probability with rainfall amount and AMC presents the suitability of the approach in decision making in an uncertain environment. The methodology has been applied to the Posina basin in Italy. From the promising results obtained, we can conclude that the Bayesian approach in flash flood forecasting provides more realistic forecasting over the FFG.
An IM-COH Algorithm Neural Network Optimization with Cuckoo Search Algorithm for Time Series Samples
Back propagation algorithm (BP) is a widely used
technique in artificial neural network and has been used as a tool
for solving the time series problems, such as decreasing training
time, maximizing the ability to fall into local minima, and optimizing
sensitivity of the initial weights and bias. This paper proposes an
improvement of a BP technique which is called IM-COH algorithm
(IM-COH). By combining IM-COH algorithm with cuckoo search
algorithm (CS), the result is cuckoo search improved control output
hidden layer algorithm (CS-IM-COH). This new algorithm has a
better ability in optimizing sensitivity of the initial weights and bias
than the original BP algorithm. In this research, the algorithm of
CS-IM-COH is compared with the original BP, the IM-COH, and the
original BP with CS (CS-BP). Furthermore, the selected benchmarks,
four time series samples, are shown in this research for illustration.
The research shows that the CS-IM-COH algorithm give the best
forecasting results compared with the selected samples.
Comparison of Machine Learning Models for the Prediction of System Marginal Price of Greek Energy Market
The Greek Energy Market is structured as a mandatory pool where the producers make their bid offers in day-ahead basis. The System Operator solves an optimization routine aiming at the minimization of the cost of produced electricity. The solution of the optimization problem leads to the calculation of the System Marginal Price (SMP). Accurate forecasts of the SMP can lead to increased profits and more efficient portfolio management from the producer`s perspective. Aim of this study is to provide a comparative analysis of various machine learning models such as artificial neural networks and neuro-fuzzy models for the prediction of the SMP of the Greek market. Machine learning algorithms are favored in predictions problems since they can capture and simulate the volatilities of complex time series.
Forecasting Issues in Energy Markets within a Reg-ARIMA Framework
Electricity markets throughout the world have
undergone substantial changes. Accurate, reliable, clear and
comprehensible modeling and forecasting of different variables
(loads and prices in the first instance) have achieved increasing
importance. In this paper, we describe the actual state of the
art focusing on reg-SARMA methods, which have proven to be
flexible enough to accommodate the electricity price/load behavior
satisfactory. More specifically, we will discuss: 1) The dichotomy
between point and interval forecasts; 2) The difficult choice between
stochastic (e.g. climatic variation) and non-deterministic predictors
(e.g. calendar variables); 3) The confrontation between modelling
a single aggregate time series or creating separated and potentially
different models of sub-series. The noteworthy point that we would
like to make it emerge is that prices and loads require different
approaches that appear irreconcilable even though must be made
reconcilable for the interests and activities of energy companies.
Load Forecasting Using Neural Network Integrated with Economic Dispatch Problem
High cost of fossil fuels and intensifying installations of alternate energy generation sources are intimidating main challenges in power systems. Making accurate load forecasting an important and challenging task for optimal energy planning and management at both distribution and generation side. There are many techniques to forecast load but each technique comes with its own limitation and requires data to accurately predict the forecast load. Artificial Neural Network (ANN) is one such technique to efficiently forecast the load. Comparison between two different ranges of input datasets has been applied to dynamic ANN technique using MATLAB Neural Network Toolbox. It has been observed that selection of input data on training of a network has significant effects on forecasted results. Day-wise input data forecasted the load accurately as compared to year-wise input data. The forecasted load is then distributed among the six generators by using the linear programming to get the optimal point of generation. The algorithm is then verified by comparing the results of each generator with their respective generation limits.
Evaluation of Best-Fit Probability Distribution for Prediction of Extreme Hydrologic Phenomena
The probability distributions are the best method for forecasting of extreme hydrologic phenomena such as rainfall and flood flows. In this research, in order to determine suitable probability distribution for estimating of annual extreme rainfall and flood flows (discharge) series with different return periods, precipitation with 40 and discharge with 58 years time period had been collected from Karkheh River at Iran. After homogeneity and adequacy tests, data have been analyzed by Stormwater Management and Design Aid (SMADA) software and residual sum of squares (R.S.S). The best probability distribution was Log Pearson Type III with R.S.S value (145.91) and value (13.67) for peak discharge and Log Pearson Type III with R.S.S values (141.08) and (8.95) for maximum discharge in Jelogir Majin and Pole Zal stations, respectively. The best distribution for maximum precipitation in Jelogir Majin and Pole Zal stations was Log Pearson Type III distribution with R.S.S values (1.74&1.90) and then Pearson Type III distribution with R.S.S values (1.53&1.69). Overall, the Log Pearson Type III distributions are acceptable distribution types for representing statistics of extreme hydrologic phenomena in Karkheh River at Iran with the Pearson Type III distribution as a potential alternative.
Influence of a High-Resolution Land Cover Classification on Air Quality Modelling
Poor air quality is one of the main environmental causes of premature deaths worldwide, and mainly in cities, where the majority of the population lives. It is a consequence of successive land cover (LC) and use changes, as a result of the intensification of human activities. Knowing these landscape modifications in a comprehensive spatiotemporal dimension is, therefore, essential for understanding variations in air pollutant concentrations. In this sense, the use of air quality models is very useful to simulate the physical and chemical processes that affect the dispersion and reaction of chemical species into the atmosphere. However, the modelling performance should always be evaluated since the resolution of the input datasets largely dictates the reliability of the air quality outcomes. Among these data, the updated LC is an important parameter to be considered in atmospheric models, since it takes into account the Earth’s surface changes due to natural and anthropic actions, and regulates the exchanges of fluxes (emissions, heat, moisture, etc.) between the soil and the air. This work aims to evaluate the performance of the Weather Research and Forecasting model coupled with Chemistry (WRF-Chem), when different LC classifications are used as an input. The influence of two LC classifications was tested: i) the 24-classes USGS (United States Geological Survey) LC database included by default in the model, and the ii) CLC (Corine Land Cover) and specific high-resolution LC data for Portugal, reclassified according to the new USGS nomenclature (33-classes). Two distinct WRF-Chem simulations were carried out to assess the influence of the LC on air quality over Europe and Portugal, as a case study, for the year 2015, using the nesting technique over three simulation domains (25 km2, 5 km2 and 1 km2 horizontal resolution). Based on the 33-classes LC approach, particular emphasis was attributed to Portugal, given the detail and higher LC spatial resolution (100 m x 100 m) than the CLC data (5000 m x 5000 m). As regards to the air quality, only the LC impacts on tropospheric ozone concentrations were evaluated, because ozone pollution episodes typically occur in Portugal, in particular during the spring/summer, and there are few research works relating to this pollutant with LC changes. The WRF-Chem results were validated by season and station typology using background measurements from the Portuguese air quality monitoring network. As expected, a better model performance was achieved in rural stations: moderate correlation (0.4 – 0.7), BIAS (10 – 21µg.m-3) and RMSE (20 – 30 µg.m-3), and where higher average ozone concentrations were estimated. Comparing both simulations, small differences grounded on the Leaf Area Index and air temperature values were found, although the high-resolution LC approach shows a slight enhancement in the model evaluation. This highlights the role of the LC on the exchange of atmospheric fluxes, and stresses the need to consider a high-resolution LC characterization combined with other detailed model inputs, such as the emission inventory, to improve air quality assessment.
PM10 Prediction and Forecasting Using CART: A Case Study for Pleven, Bulgaria
Ambient air pollution with fine particulate matter (PM10) is a systematic permanent problem in many countries around the world. The accumulation of a large number of measurements of both the PM10 concentrations and the accompanying atmospheric factors allow for their statistical modeling to detect dependencies and forecast future pollution. This study applies the classification and regression trees (CART) method for building and analyzing PM10 models. In the empirical study, average daily air data for the city of Pleven, Bulgaria for a period of 5 years are used. Predictors in the models are seven meteorological variables, time variables, as well as lagged PM10 variables and some lagged meteorological variables, delayed by 1 or 2 days with respect to the initial time series, respectively. The degree of influence of the predictors in the models is determined. The selected best CART models are used to forecast future PM10 concentrations for two days ahead after the last date in the modeling procedure and show very accurate results.
Statistical and Land Planning Study of Tourist Arrivals in Greece during 2005-2016
During the last 10 years, in spite of the economic crisis, the number of tourists arriving in Greece has increased, particularly during the tourist season from April to October. In this paper, the number of annual tourist arrivals is studied to explore their preferences with regard to the month of travel, the selected destinations, as well the amount of money spent. The collected data are processed with statistical methods, yielding numerical and graphical results. From the computation of statistical parameters and the forecasting with exponential smoothing, useful conclusions are arrived at that can be used by the Greek tourism authorities, as well as by tourist organizations, for planning purposes for the coming years. The results of this paper and the computed forecast can also be used for decision making by private tourist enterprises that are investing in Greece. With regard to the statistical methods, the method of Simple Exponential Smoothing of time series of data is employed. The search for a best forecast for 2017 and 2018 provides the value of the smoothing coefficient. For all statistical computations and graphics Microsoft Excel is used.
Real Time Classification of Political Tendency of Twitter Spanish Users based on Sentiment Analysis
What people say on social media has turned into a
rich source of information to understand social behavior. Specifically,
the growing use of Twitter social media for political communication
has arisen high opportunities to know the opinion of large numbers
of politically active individuals in real time and predict the global
political tendencies of a specific country. It has led to an increasing
body of research on this topic. The majority of these studies have
been focused on polarized political contexts characterized by only
two alternatives. Unlike them, this paper tackles the challenge
of forecasting Spanish political trends, characterized by multiple
political parties, by means of analyzing the Twitters Users political
tendency. According to this, a new strategy, named Tweets Analysis
Strategy (TAS), is proposed. This is based on analyzing the users
tweets by means of discovering its sentiment (positive, negative or
neutral) and classifying them according to the political party they
support. From this individual political tendency, the global political
prediction for each political party is calculated. In order to do this,
two different strategies for analyzing the sentiment analysis are
proposed: one is based on Positive and Negative words Matching
(PNM) and the second one is based on a Neural Networks Strategy
(NNS). The complete TAS strategy has been performed in a Big-Data
environment. The experimental results presented in this paper reveal
that NNS strategy performs much better than PNM strategy to analyze
the tweet sentiment. In addition, this research analyzes the viability
of the TAS strategy to obtain the global trend in a political context
make up by multiple parties with an error lower than 23%.
Deep Learning for Renewable Power Forecasting: An Approach Using LSTM Neural Networks
Load forecasting has become crucial in recent years
and become popular in forecasting area. Many different power
forecasting models have been tried out for this purpose. Electricity
load forecasting is necessary for energy policies, healthy and reliable
grid systems. Effective power forecasting of renewable energy load
leads the decision makers to minimize the costs of electric utilities
and power plants. Forecasting tools are required that can be used
to predict how much renewable energy can be utilized. The purpose
of this study is to explore the effectiveness of LSTM-based neural
networks for estimating renewable energy loads. In this study, we
present models for predicting renewable energy loads based on
deep neural networks, especially the Long Term Memory (LSTM)
algorithms. Deep learning allows multiple layers of models to learn
representation of data. LSTM algorithms are able to store information
for long periods of time. Deep learning models have recently been
used to forecast the renewable energy sources such as predicting
wind and solar energy power. Historical load and weather information
represent the most important variables for the inputs within the
power forecasting models. The dataset contained power consumption
measurements are gathered between January 2016 and December
2017 with one-hour resolution. Models use publicly available data
from the Turkish Renewable Energy Resources Support Mechanism.
Forecasting studies have been carried out with these data via deep
neural networks approach including LSTM technique for Turkish
electricity markets. 432 different models are created by changing
layers cell count and dropout. The adaptive moment estimation
(ADAM) algorithm is used for training as a gradient-based optimizer
instead of SGD (stochastic gradient). ADAM performed better than
SGD in terms of faster convergence and lower error rates. Models
performance is compared according to MAE (Mean Absolute Error)
and MSE (Mean Squared Error). Best five MAE results out of
432 tested models are 0.66, 0.74, 0.85 and 1.09. The forecasting
performance of the proposed LSTM models gives successful results
compared to literature searches.
Electricity Price Forecasting: A Comparative Analysis with Shallow-ANN and DNN
Electricity prices have sophisticated features such as
high volatility, nonlinearity and high frequency that make forecasting
quite difficult. Electricity price has a volatile and non-random
character so that, it is possible to identify the patterns based on the
historical data. Intelligent decision-making requires accurate price
forecasting for market traders, retailers, and generation companies.
So far, many shallow-ANN (artificial neural networks) models have
been published in the literature and showed adequate forecasting
results. During the last years, neural networks with many hidden
layers, which are referred to as DNN (deep neural networks) have
been using in the machine learning community. The goal of this
study is to investigate electricity price forecasting performance of the
shallow-ANN and DNN models for the Turkish day-ahead electricity
market. The forecasting accuracy of the models has been evaluated
with publicly available data from the Turkish day-ahead electricity
market. Both shallow-ANN and DNN approach would give successful
result in forecasting problems. Historical load, price and weather
temperature data are used as the input variables for the models.
The data set includes power consumption measurements gathered
between January 2016 and December 2017 with one-hour resolution.
In this regard, forecasting studies have been carried out comparatively
with shallow-ANN and DNN models for Turkish electricity markets
in the related time period. The main contribution of this study
is the investigation of different shallow-ANN and DNN models
in the field of electricity price forecast. All models are compared
regarding their MAE (Mean Absolute Error) and MSE (Mean Square)
results. DNN models give better forecasting performance compare to
shallow-ANN. Best five MAE results for DNN models are 0.346,
0.372, 0.392, 0,402 and 0.409.
Time Series Modelling and Prediction of River Runoff: Case Study of Karkheh River, Iran
Rainfall and runoff phenomenon is a chaotic and complex outcome of nature which requires sophisticated modelling and simulation methods for explanation and use. Time Series modelling allows runoff data analysis and can be used as forecasting tool. In the paper attempt is made to model river runoff data and predict the future behavioural pattern of river based on annual past observations of annual river runoff. The river runoff analysis and predict are done using ARIMA model. For evaluating the efficiency of prediction to hydrological events such as rainfall, runoff and etc., we use the statistical formulae applicable. The good agreement between predicted and observation river runoff coefficient of determination (R2) display that the ARIMA (4,1,1) is the suitable model for predicting Karkheh River runoff at Iran.
A Non-Linear Eddy Viscosity Model for Turbulent Natural Convection in Geophysical Flows
Eddy viscosity models in turbulence modeling can be mainly classified as linear and nonlinear models. Linear formulations are simple and require less computational resources but have the disadvantage that they cannot predict actual flow pattern in complex geophysical flows where streamline curvature and swirling motion are predominant. A constitutive equation of Reynolds stress anisotropy is adopted for the formulation of eddy viscosity including all the possible higher order terms quadratic in the mean velocity gradients, and a simplified model is developed for actual oceanic flows where only the vertical velocity gradients are important. The new model is incorporated into the one dimensional General Ocean Turbulence Model (GOTM). Two realistic oceanic test cases (OWS Papa and FLEX' 76) have been investigated. The new model predictions match well with the observational data and are better in comparison to the predictions of the two equation k-epsilon model. The proposed model can be easily incorporated in the three dimensional Princeton Ocean Model (POM) to simulate a wide range of oceanic processes. Practically, this model can be implemented in the coastal regions where trasverse shear induces higher vorticity, and for prediction of flow in estuaries and lakes, where depth is comparatively less. The model predictions of marine turbulence and other related data (e.g. Sea surface temperature, Surface heat flux and vertical temperature profile) can be utilized in short term ocean and climate forecasting and warning systems.
Flood Predicting in Karkheh River Basin Using Stochastic ARIMA Model
Floods have huge environmental and economic impact. Therefore, flood prediction is given a lot of attention due to its importance. This study analysed the annual maximum streamflow (discharge) (AMS or AMD) of Karkheh River in Karkheh River Basin for flood predicting using ARIMA model. For this purpose, we use the Box-Jenkins approach, which contains four-stage method model identification, parameter estimation, diagnostic checking and forecasting (predicting). The main tool used in ARIMA modelling was the SAS and SPSS software. Model identification was done by visual inspection on the ACF and PACF. SAS software computed the model parameters using the ML, CLS and ULS methods. The diagnostic checking tests, AIC criterion, RACF graph and RPACF graphs, were used for selected model verification. In this study, the best ARIMA models for Annual Maximum Discharge (AMD) time series was (4,1,1) with their AIC value of 88.87. The RACF and RPACF showed residuals’ independence. To forecast AMD for 10 future years, this model showed the ability of the model to predict floods of the river under study in the Karkheh River Basin. Model accuracy was checked by comparing the predicted and observation series by using coefficient of determination (R2).
Forecasting of Scaffolding Work Comfort Parameters Based on Data from Meteorological Stations
Work at height, such as construction works on scaffoldings, is associated with a considerable risk. Scaffolding workers are usually exposed to changing weather conditions what can additionally increase the risk of dangerous situations. Therefore, it is very important to foresee the risk of adverse conditions to which the worker may be exposed. The data from meteorological stations may be used to asses this risk. However, the dependency between weather conditions on a scaffolding and in the vicinity of meteorological station, should be determined. The paper presents an analysis of two selected environmental parameters which have influence on the behavior of workers – air temperature and wind speed. Measurements of these parameters were made between April and November of 2016 on ten scaffoldings located in different parts of Poland. They were compared with the results taken from the meteorological stations located closest to the studied scaffolding. The results gathered from the construction sites and meteorological stations were not the same, but statistical analyses have shown that they were correlated.
A Comparative Analysis of the Performance of COSMO and WRF Models in Quantitative Rainfall Prediction
The Numerical weather prediction (NWP) models are
considered powerful tools for guiding quantitative rainfall prediction.
A couple of NWP models exist and are used at many operational
weather prediction centers. This study considers two models namely
the Consortium for Small–scale Modeling (COSMO) model and the
Weather Research and Forecasting (WRF) model. It compares the
models’ ability to predict rainfall over Uganda for the period 21st
April 2013 to 10th May 2013 using the root mean square (RMSE)
and the mean error (ME). In comparing the performance of the
models, this study assesses their ability to predict light rainfall events
and extreme rainfall events. All the experiments used the default
parameterization configurations and with same horizontal resolution
(7 Km). The results show that COSMO model had a tendency of
largely predicting no rain which explained its under–prediction. The
COSMO model (RMSE: 14.16; ME: -5.91) presented a significantly
(p = 0.014) higher magnitude of error compared to the WRF
model (RMSE: 11.86; ME: -1.09). However the COSMO model
(RMSE: 3.85; ME: 1.39) performed significantly (p = 0.003) better
than the WRF model (RMSE: 8.14; ME: 5.30) in simulating light
rainfall events. All the models under–predicted extreme rainfall events
with the COSMO model (RMSE: 43.63; ME: -39.58) presenting
significantly higher error magnitudes than the WRF model (RMSE:
35.14; ME: -26.95). This study recommends additional diagnosis of
the models’ treatment of deep convection over the tropics.
A Comparative Analysis of Artificial Neural Network and Autoregressive Integrated Moving Average Model on Modeling and Forecasting Exchange Rate
This paper examines the forecasting performance of Autoregressive Integrated Moving Average (ARIMA) and Artificial Neural Networks (ANN) models with the published exchange rate obtained from South African Reserve Bank (SARB). ARIMA is one of the popular linear models in time series forecasting for the past decades. ARIMA and ANN models are often compared and literature revealed mixed results in terms of forecasting performance. The study used the MSE and MAE to measure the forecasting performance of the models. The empirical results obtained reveal the superiority of ARIMA model over ANN model. The findings further resolve and clarify the contradiction reported in literature over the superiority of ARIMA and ANN models.
Input Data Balancing in a Neural Network PM-10 Forecasting System
Recently PM-10 has become a social and global issue. It is one of major air pollutants which affect human health. Therefore, it needs to be forecasted rapidly and precisely. However, PM-10 comes from various emission sources, and its level of concentration is largely dependent on meteorological and geographical factors of local and global region, so the forecasting of PM-10 concentration is very difficult. Neural network model can be used in the case. But, there are few cases of high concentration PM-10. It makes the learning of the neural network model difficult. In this paper, we suggest a simple input balancing method when the data distribution is uneven. It is based on the probability of appearance of the data. Experimental results show that the input balancing makes the neural networks’ learning easy and improves the forecasting rates.
Forecasting the Volatility of Geophysical Time Series with Stochastic Volatility Models
This work is devoted to the study of modeling
geophysical time series. A stochastic technique with time-varying
parameters is used to forecast the volatility of data arising in
geophysics. In this study, the volatility is defined as a logarithmic
first-order autoregressive process. We observe that the inclusion of
log-volatility into the time-varying parameter estimation significantly
improves forecasting which is facilitated via maximum likelihood
estimation. This allows us to conclude that the estimation algorithm
for the corresponding one-step-ahead suggested volatility (with ±2
standard prediction errors) is very feasible since it possesses good
Sphere in Cube Grid Approach to Modelling of Shale Gas Production Using Non-Linear Flow Mechanisms
Shale gas is one of the most rapidly growing forms of natural gas. Unconventional natural gas deposits are difficult to characterize overall, but in general are often lower in resource concentration and dispersed over large areas. Moreover, gas is densely packed into the matrix through adsorption which accounts for large volume of gas reserves. Gas production from tight shale deposits are made possible by extensive and deep well fracturing which contacts large fractions of the formation. The conventional reservoir modelling and production forecasting methods, which rely on fluid-flow processes dominated by viscous forces, have proved to be very pessimistic and inaccurate. This paper presents a new approach to forecast shale gas production by detailed modeling of gas desorption, diffusion and non-linear flow mechanisms in combination with statistical representation of these processes. The representation of the model involves a cube as a porous media where free gas is present and a sphere (SiC: Sphere in Cube model) inside it where gas is adsorbed on to the kerogen or organic matter. Further, the sphere is considered consisting of many layers of adsorbed gas in an onion-like structure. With pressure decline, the gas desorbs first from the outer most layer of sphere causing decrease in its molecular concentration. The new available surface area and change in concentration triggers the diffusion of gas from kerogen. The process continues until all the gas present internally diffuses out of the kerogen, gets adsorbs onto available surface area and then desorbs into the nanopores and micro-fractures in the cube. Each SiC idealizes a gas pathway and is characterized by sphere diameter and length of the cube. The diameter allows to model gas storage, diffusion and desorption; the cube length takes into account the pathway for flow in nanopores and micro-fractures. Many of these representative but general cells of the reservoir are put together and linked to a well or hydraulic fracture. The paper quantitatively describes these processes as well as clarifies the geological conditions under which a successful shale gas production could be expected. A numerical model has been derived which is then compiled on FORTRAN to develop a simulator for the production of shale gas by considering the spheres as a source term in each of the grid blocks. By applying SiC to field data, we demonstrate that the model provides an effective way to quickly access gas production rates from shale formations. We also examine the effect of model input properties on gas production.