Aims Press

Research article

AIMS Environmental Science, doi: 10.3934/environsci.2015.2.122

Export file:

Format

Content

The role of lipids in activated sludge floc formation

School of Biotechnology and Biomolecular Sciences, University of NSW, Sydney 2052, Australia

Activated sludge is widely used to treat municipal and industrial wastewater globally and the formation of activated sludge flocculates (flocs) underpins the ability to separate sludge from treated water. Despite the importance of activated sludge flocs to human civilization there have been precious few attempts to rationally design fit for purpose flocs using a bottom-up approach based on a solid scientific foundation. Recently we have been developing experimental models for activated sludge floc formation based on the colonization and consumption of particulate organic matter (chitin and cellulose). In this study we lay the foundation for investigation of activated sludge floc formation based on biofilm formation around spheres of the lipid glycerol trioleate (GT) that form spontaneously when GT is introduced into activated sludge incubations. Sludge biomass was observed to associate tightly with the lipid spheres. An increase in extracellular lipase activity was associated with a decrease in size of the colonized lipid spheres over a 25 day incubation. Bacterial community composition shifted from predominantly Betaproteobacteria to Alphaproteobacteria in GT treated sludge. Four activated sludge bacteria were isolated from lipid spheres and two of them were shown to produce AHL like quorum sensing signal activity, suggesting quorum sensing may play a role in lipid spheres colonization and biodegradation in activated sludge. The development of this experimental model of activated sludge floc formation lays the foundation for rational production of flocs for wastewater treatment using lipids as floc nuclei and further development of the flocculate life-cycle concept.

1. Introduction

Measuring volatility is of great importance in empirical finance to manage risk, such as portfolio allocation and to derivative asset pricing. Since volatility is not directly observable in financial markets, we must rely on some measurement techniques that estimate volatility. One of the measurement techniques, presumably the most popular, is the autoregressive conditional heteroscedasticity (ARCH) model (Engle, 1982) and its generalized version, the generalized autoregressive conditional heteroscedasticity (GARCH) model (Bollerslev, 1986). These "ARCH-type" models can successfully capture some of stylized facts of asset returns such as volatility clustering and the fat-tailed nature of return distributions. Researchers proposed numerous extensions to the ARCH or GARCH models designed to capture more features of asset returns and with the expectation of estimating more accurate volatilities. The asymmetric GARCH models such as the exponential GARCH (EGARCH) (Nelson, 1991) and GJR-GARCH (Glosten, 1993) models aim to capture the asymmetric property of the volatility process. Certain assets, such as stock prices, display asymmetric volatility. It is crucial to introduce the asymmetric volatility process in the model to obtain more accurate estimates for these assets. The rational GARCH (RGARCH) model (Takaishi, 2017) is another asymmetric model that uses a rational form in the volatility process. Most volatility processes of ARCH-type models are constructed as a polynomial of past volatilities and returns. The polynomial can be viewed as a Taylor expansion of the yet unknown true volatility process (Sentana, 1995) and the GARCH (1, 1) model corresponds to the lowest order Taylor expansion for the symmetric volatility process. The RGARCH model is based on a rational polynomial or "Padé approximants, " which could approximate the true volatility process better than Taylor polynomials do. Various fields use rational polynomials to obtain a better approximate functions. For instance, in lattice quantum chromodynamics simulations, the power of a fermion matrix is approximated by a rational function, which is used in the hybrid Monte Carlo method (Clark, 2006, 2007). This then shows that the rational form is superior to the polynomial version of the hybrid Monte Carlo method (de Forcrand, 1997; Takaishi, 2001b, 2001c, 2001a, 2002). In finance, some studies applied Padé approximants to describe return probability distributions (Nuyts, 2001; Alderweireld, 2004) and error terms of the GARCH model (Takaishi, 2012; Chen, 2013). The RGARCH model has been examined for stock returns by the goodness of fit test with the deviance information criterion (DIC) (Spiegelhalter, 2002), and some claim that the RGARCH model can be equally effective with other asymmetric GARCH models. In this study, we further examine the performance of volatility estimation with the RGARCH model. We compare the accuracy of volatility estimations by a loss function using realized volatility (RV) (Andersen, 1998; Barndorff-Nielsen, 2001) as a proxy of true volatility.

The rest of this paper is organized as follows. Section 2 introduces the RGARCH model and the other GARCH-type models we use in this study. In Section 3, we describe the Bayesian inference we use for the parameter estimation of the models and in Section 4, we present the comparative accuracy results of the volatility estimations. Finally, we conclude in Section 5.


2. Rational GARCH model

The RGARCH model captures volatility asymmetry in a rational form as

$ \sigma_t^2 = \frac{\omega + \alpha r_{t-1}^2 + \beta \sigma_{t-1}^2}{1+\delta r_{t-1}}, $ (1)

where $\sigma_t^2$ and $r_t$ is volatility and return at time $t$, respectively, and $\alpha, \beta, \omega$ and $\delta$ are the model parameters. $\delta$ introduces asymmetry against $r_{t-1}$ in the volatility process. If $\delta$ is positive, then the volatility will be higher for negative returns. For $\delta = 0$, the RGARCH model reduces to the well-known symmetric model, that is, the GARCH (1, 1) model. The return $r_t$ at time $t$ is defined by $r_t = \sigma_t \epsilon_t$, where $\epsilon_t \sim N(0, 1)$.

Equation (1) is not well-defined when the denominator of Equation (1) is negative. The denominator will be negative for $1 < -\delta r_{t-1}$. Although we did not experience a negative denominator in this study, it may occur for huge $r_{t-1}$ or outliers. To avoid such cases, we also propose an alternative form:

$ \sigma_t^2 = \frac{\omega + \alpha r_{t-1}^2 + \beta \sigma_{t-1}^2}{\exp(\delta r_{t-1})}. $ (2)

We call this model the "RGARCH-Exp" model. For a small $\delta r_{t-1}$, equation (2) reduces to equation (1).

To compare the accuracy of the volatility estimation, we also consider the following GARCH-type models.

● GARCH model (Bollerslev, 1986)

$ \sigma_t^2 = \omega + \alpha r_{t-1}^2 + \beta \sigma_{t-1}^2. $ (3)

● Exponential GARCH (EGARCH) model (Nelson, 1991)

$ \ln \sigma_t^2 = \omega + \beta \ln \sigma_{t-1}^2 + \theta z_{t-1} +\gamma (|z_{t-1}| -E(|z_{t-1}|)), $ (4)

where $z_{t-1} = r_{t-1}/\sigma_{t-1}$.

● GJR model (Glosten, 1993)

$ \sigma^2_t = \left\{ \begin{array}{ll} \omega + \alpha r_{t-1}^2 + \beta \sigma_{t-1}^2, & r_{t-1} \ge 0\\ \omega + (\alpha+\rho) r_{t-1}^2 + \beta \sigma_{t-1}^2, & r_{t-1} < 0. \end{array} \right. $ (5)

The GARCH model is the symmetric volatility model and the EGARCH and GJR models correspond to the asymmetric volatility models.


3. Bayesian inference

In this study, we employ Bayesian inference for the parameter estimation of the models. From the Bayes theorem, we obtain the posterior density $P(\theta|r)$ with $T$ observation $r = (r_1, r_2, \dots, r_T)$ given by

$ P(\theta|r) \propto L(r|\theta)\pi(\theta), $ (6)

where $L(y|\theta)$ represents the likelihood function and $\pi(\theta)$ is the prior density for $\theta$. Here, $\theta$ represents model parameters such as $\theta = (\alpha, \beta, \omega)$ for the GARCH model. We assume that the prior density is constant. The likelihood function is given by

$ L(r|\theta) = \prod\limits_{i = 1}^T (2\pi\sigma_t^2)^{1/2} \exp\left(-\frac{r_i^2}{\sigma_t^2}\right). $ (7)

Using the posterior density $P(\theta|r)$, we infer the model parameters as the expectation values by

$ \left\langle \theta \right\rangle = \frac1Z \int \theta P(\theta|r)d\theta, $ (8)

where $Z = \int P(\theta|r)d\theta$ is the normalization constant.

Generally, we cannot solve equation (8) analytically. Hence, we estimate equation (8) by the Markov Chain Monte Carlo (MCMC) method using the Metropolis-Hastings (MH) algorithm (Hastings, 1970). We generate samples from the probability distribution $P(\theta|r)$. In the MH algorithm, starting from $\theta$, we first propose a candidate $\theta^\prime$ from a certain proposal density $g(\theta^\prime|\theta)$. Then, we accept the candidate $\theta^\prime$ with a probability of

$ P_{MH}(\theta^\prime|\theta) = \min \left[ 1, \frac{P(\theta^\prime|r)g(\theta|\theta^\prime)}{P(\theta|r)g(\theta^\prime|\theta)}\right]. $ (9)

If the candidate $\theta^\prime$ is rejected, we keep $\theta$.

Following Refs.(Mitsui, 2003; Asai, 2006; Takaishi, 2009a), we use a multivariate Student t-distribution for the proposal density. The $p$-dimensional multivariate Student t-distribution is given by

$ \begin{eqnarray*} g(\theta|\theta^\prime) & = & \frac{\Gamma((\nu+p)/2)/\Gamma(\nu/2)}{\det \Sigma^{1/2} (\nu\pi)^{p/2}} \nonumber \\ & & \times \left[1+\frac{(\theta-M)^t \Sigma^{-1}(\theta-M)}{\nu}\right]^{-(\nu+p)/2}, \end{eqnarray*} $ (10)

where $\theta$ and $M$ are column vectors,

$ \theta = \left[ \begin{array}{c} \theta_1 \\ \theta_2 \\ \vdots \\ \theta_p \end{array} \right], M = \left[ \begin{array}{c} M_1 \\ M_2 \\ \vdots \\ M_p \end{array} \right], $ (11)
$ M_i = E(\theta_i), $ (12)

and $p$ corresponds to the number of model parameters. $\Sigma$ is the covariance matrix defined as

$ \frac{\nu\Sigma}{\nu-2} = E[(\theta-M)(\theta-M)^t]. $ (13)

$\nu$ is a parameter to tune the shape of the Student t-distribution. In this study, we choose $\nu = 10$ (Takaishi, 2009a). Note that since equation (10) does not depend on the previous value of $\theta^\prime$, we obtain $g(\theta|\theta^\prime) = g(\theta)$.

To utilize equation (10) in the MH algorithm, we need to specify the unknown parameters $M$ and $\Sigma$. We determine $M$ and $\Sigma$ through MCMC simulations(Takaishi, 2009a, 2009b, 2009c, 2010). First, we make a short pilot run using the Metropolis algorithm (Metropolis, 1953) and accumulate some data. We then estimate $M$ and $\Sigma$ by equations (12) and (13). Second, substituting the estimated $M$ and $\Sigma$ into eq.(10), we perform the MH simulation. After accumulating more sample data, we recalculate $M$ and $\Sigma$, and update $M$ and $\Sigma$ in equation (10).


4. Empirical results

We use daily closing stock price time series data from the Tokyo Stock Exchange from January 5,2004 to December 30,2015 for the following six stocks: 1 Japan Tobacco (JT) Inc., 2 Canon Inc., 3 Toray Industries Inc., 4 Panasonic Co., 5 Sumitomo Metal Industries Ltd., and 6 Nippon Steel Co. The daily return $R_t$ is defined by $100\times (ln P(t+1)-ln P(t))$, where $P(t)$ is the daily closing price at day $t$. For the RV calculations, we use 1-min high frequency return data for the same period. We evaluate the performance of the volatility estimate compared to the RV as a proxy of true volatility. To be specific, we calculate the following QLIKE loss function (Patton, 2011):

$ QLIKE(\Delta) = \frac1T\sum\limits_{t = 1}^T (\ln \sigma_t^2 -\frac{RV_{t, \Delta}}{\sigma_t^2}), $ (14)

where $RV_{t, \Delta}$ is the RV at time $t$ calculated at the sampling frequency $\Delta$. The $RV_{t, \Delta}$ is defined by the sum of the squared intraday returns,

$ RV_{t, \Delta} = \sum\limits_{i = 1}^{n(\Delta)} r_{t+i\Delta, \Delta}^2. $ (15)

where $r_{t, \Delta}$ is an intraday return sampled at the frequency of $\Delta$ at $t$ and $n(\Delta)$ is the number of intraday returns in a day. We use $RV_{t, \Delta}$ at $\Delta = 1, 2, ..., 40$. Thus, there are 40 RVs in total at various sampling frequencies. In ideal circumstances, equation (15) goes to the true volatility in the limit of $\Delta \rightarrow 0$. However, in practice, due to the microstructure noise and non-trading hours, the actual RV is biased. We modify the RV using the bias correction factor introduced by Hansen and Lunde (Hansen, 2005) given by

$ C_\Delta = \frac{\sum\limits_{i = 1}^T (R_t^D-\bar{R^D})^2}{\sum\limits_{i = 1}^T RV_{t, \Delta}}, $ (16)

where $R_t^D$ is the daily return, $\bar{R^D}$ is the average of the daily returns, and $T$ is the number of daily returns. Using $C_\Delta$, the modified RV, we obtain $RV_{t, \Delta}^\prime$ using $RV_{t, \Delta}^\prime = C_\Delta RV_{t, \Delta}$.

We estimate the model parameters with the Bayesian inference described in Section 3. We perform the Bayesian inference using the MH algorithm with a multi-dimensional Student t proposal density. First, we estimate the parameters of the proposal density with the Metropolis algorithm as a pilot run, and then switch to the MH algorithm. We recalculate the parameters of the proposal density every 1,000 Monte Carlo (MC) updates using all of the MC data. After the 5000 repetition burn-in process, we collect 30,000 data points for analysis. Figure 1 shows the acceptance of the MH algorithm every 1,000 MC updates for the RGARCH-Exp model with JT stock data. The initial acceptance is low since the parameters of the proposal density are still not estimated accurately. Then, as the MC run proceeds, the acceptance increases quickly to about 70%. We obtain similar results for other cases.

Figure 1. MH algorithm acceptance for the RGARCH-Exp model with JT stock data as a function of the MC update..

Figure 2 represents daily return time series of JT, the volatility $\sigma_t^2$ estimated by the MCMC simulation with the RGARCH-Exp model and the modified RV at $\Delta = 1min$ as representative. The estimated volatility is obtained by an average over the MC samples and its accuracy compared to the RV is examined by the loss function.

Figure 2. Return time series, volatility estimated using the MCMC simulation with the RGARCH-Exp model and the modified RV at $\Delta = 1min$.

Table 1 summarizes the results for the model parameters and the QLIKE loss function. We provide the QLIKE values as an average over 40 values of $QLIKE(\Delta)$ at $\Delta = 1, 2, ..., 40$. The values in square parentheses indicate the ranking based on QLIKE. Since the ranking varies depending on the stock price, we also give an average over six stocks in square parentheses in the first column at QLIKE. The average QLIKEs for all asymmetric GARCH-type models are less than that of the GARCH model, which indicates that for the stock price time series, introducing asymmetry in the volatility process is crucial to obtaining accurate volatilities. According to the average QLIKE, the RGARCH-Exp model should rank first. However, for individual stocks, the other symmetric GARCH-type models can take the rank first. For instance, the RGARCH, EGARCH, and GJR models rank first for Toray, Canon, and, Panasonic respectively. Thus, we conclude that all asymmetric GARCH-type models examined in this study perform equally well. In Table 1, we also provide the DIC (Spiegelhalter, 2002) which measures the goodness-of-fit of the models. The values in square parentheses indicate the ranking results based on DIC. We find that the DIC results show the similar ranking to the QLIKE loss function. Namley all asymmetric GARCH-type models are superior to the GARCH model and the RGARCH-Exp model ranks first.

Table 1. Model parameters and QLIKE results. Values in round parentheses denote the standard deviation of the data sampled with the MCMC method. The values in square parentheses indicate the ranking results. The values in square parentheses in the first column show averages of ranking over six stocks.
RGARCH JT Canon Toray Panasonic Sumitomo Nippon
$\alpha$ 0.095(21) 0.099(22) 0.091(22) 0.127(30) 0.091(22) 0.142(28)
$\beta$ 0.838(31) 0.881(24) 0.891(27) 0.854(31) 0.867(31) 0.818(33)
$\omega$ 0.35(11) 0.109(44) 0.089(49) 0.117(49) 0.28(12) 0.28(11)
$\delta$ 0.033(8) 0.023(6) 0.017(8) 0.022(8) 0.026(6) 0.028(8)
QLIKE [2.50] 2.7808 [2] 2.8542 [4] 2.4777 [1] 2.7213 [3] 3.1344 [1] 3.0084 [4]
DIC [2.83] 3860.89 [2] 3843.59 [4] 3680.66 [2] 3727.13 [4] 4134.67 [3] 4015.94 [2]
RGARCH-Exp
$\alpha$ 0.100(22) 0.096(22) 0.090(22) 0.123(30) 0.088(21) 0.144(28)
$\beta$ 0.847(30) 0.889(23) 0.894(27) 0.861(31) 0.880(27) 0.825(32)
$\omega$ 0.31(10) 0.094(4) 0.084(47) 0.105(47) 0.230(97) 0.252(98)
$\delta$ 0.0384(96) 0.028(8) 0.019(9) 0.028(10) 0.032(5) 0.034(9)
QLIKE [1.83] 2.7798 [1] 2.8509 [2] 2.4779 [2] 2.7209 [2] 3.1345 [2] 3.0058 [2]
DIC [1.67] 3860.16 [1] 3841.63 [3] 3680.08 [1] 3725.60[3] 4131.31 [1] 4014.06 [1]
EGARCH
$\gamma$ 0.209(39) 0.167(38) 0.198(34) 0.189(47) 0.187(35) 0.307(47)
$\beta$ 0.951(17) 0.9826(67) 0.978(89) 0.984(78) 0.969(11) 0.952(15)
$\omega$ 0.089(29) 0.029(12) 0.037(15) 0.026(14) 0.063(22) 0.092(30)
$\delta$ -0.079(23) -0.89(19) -0.028(20) -0.076(22) -0.079(20) -0.071(25)
QLIKE [3.00] 2.7822 [3] 2.8487 [1] 2.4843 [5] 2.7242 [5] 3.1347 [3] 2.9998 [1]
DIC [2.50] 3872.42 [4] 3833.53 [1] 3683.09 [3] 3714.88 [1] 4133.73 [2] 4021.71 [4]
GJR
$\alpha$ 0.053(22) 0.035(21) 0.084(23) 0.042(29) 0.052(22) 0.102(30)
$\beta$ 0.838(30) 0.897(21) 0.883(25) 0.890(29) 0.871(30) 0.808(34)
$\omega$ 0.32(10) 0.075(34) 0.091(43) 0.066(36) 0.23(11) 0.27(10)
$\rho$ 0.126(37) 0.123(29) 0.037(27) 0.123(33) 0.107(34) 0.130(48)
QLIKE [3.17] 2.7843 [5] 2.8540 [3] 2.4795 [3] 2.7188 [1] 3.1384 [4] 3.0073 [3]
DIC [3.17] 3863.33 [3] 3835.64 [2] 3683.77 [5] 3720.09 [2] 4139.65 [4] 4020.40 [3]
GARCH
$\alpha$ 0.116(22) 0.121(21) 0.100(22) 0.130(37) 0.114(24) 0.155(28)
$\beta$ 0.839(30) 0.866(25) 0.885(26) 0.854(44) 0.863(31) 0.824(31)
$\omega$ 0.30(11) 0.116(24) 0.093(46) 0.113(76) 0.23(11) 0.218(90)
QLIKE [4.50] 2.7842 [4] 2.8615 [5] 2.4802 [4] 2.7217 [4] 3.1410 [5] 3.0162 [5]
DIC [4.83] 3875.20 [5] 3852.15 [5] 3683.10 [4] 3735.91 [5] 4151.33 [5] 4026.62 [5]

5. Conclusions

We introduce the RGARCH model as an alternative GARCH model that can capture the asymmetric property of volatility. In addition to the RGARCH model, we propose a model called the RGARCH-Exp model that is more stable for outliers. We examine the performance of the volatility estimation of the RGARCH and RGARCH-Exp models with the QLIKE loss function using realized volatility to proxy for true volatility. We compare the performance of these models with that of the EGARCH, GJR, and GARCH models. We conduct empirical analyses for stock price time series for six stocks listed on the Tokyo Stock Exchange. Based on the QLIKE loss function, we find that the RGARCH, RGARCH-Exp, and other asymmetric GARCH-type models have superior performance in terms of volatility estimation compared to the GARCH model. This indicates that it is crucial to include the asymmetric volatility process in a model for accurate volatility estimations. Since our ranking varies depending on individual stock prices, and all of the asymmetric volatility models we use in this study rank first for different stocks, we conclude that performance of the RGARCH and RGARCH-Exp models are comparable to that of the asymmetric GARCH-type models.


Acknowledgment

Numerical calculations in this work were conducted at the Yukawa Institute Computer Facility and the facilities of the Institute of Statistical Mathematics.


Conflict of interest

All authors declare no conflict of interest in this paper.


  Supplementary
  Article Metrics

 

Download full text in PDF

Export Citation

Share this paper on

Article outline

Show full outline
Copyright © AIMS Press All Rights Reserved