The paper describes two algorithms for financial portfolio optimization with the following risk measures: CVaR, MAD, LSAD and dispersion CVaR. These algorithms can be applied to discrete distributions of asset returns since then the optimization problems can be reduced to linear programs. The first algorithm solves a simple recourse problem as described by Haneveld using Benders decomposition method. The second algorithm finds an optimal portfolio with the smallest distance to a given benchmark portfolio and is an adaptation of the least norm solution (called also normal solution) of linear programs due to Zhao and Li. The algorithms are implemented in R in the package PortfolioOptim.
The construction of an optimal portfolio of financial instruments is one of primal goals in asset management. Recent advances in risk measurement advocate using risk measures that take into account the tail behavior of asset return distributions, such as conditional value-at-risk or lower semi-absolute deviation. For these risk measures, finding optimal portfolios in risk-return setting leads to convex programming problems. For a finite discrete distribution of returns and many practically used risk measures the optimization problem can be solved by LP methods. The portfolio optimization by LP methods leads to two problems: for distributions with a large number of values (for example, a large sample from a continuous distribution) one obtains a unique solution but the LP problem is resource demanding; for distributions with a small number of values one often obtains a non-unique solution, which is of limited use for asset management. In the latter case, a unique optimal portfolio is obtained by projecting a benchmark portfolio on the whole set of solutions. In the paper we present an algorithm that solves efficiently the problem with large distributions and an algorithm that finds an orthogonal projection of the benchmark portfolio on the space of solutions.
The R language and environment for statistical computing offer a large variety of tools for portfolio optimization. General purpose optimization tools are reviewed by (Theussl and Borchers 2016) (R packages for solving optimization problems) and (Koenker and Mizera 2014) (R packages for convex optimization). The book by (Pfaff 2016) provides an overview of specific functions for portfolio optimization which are embedded in financial packages. Here we mention a selection of R-packages dedicated primarily to portfolio optimization. The package fPortfolio by (Würtz et al. 2009) offers a large set of functions for financial data analysis and enables portfolio optimization in mean-variance, mean-MAD and mean-CVaR settings. For these portfolio problems the package employs existing optimization tools: LP, QP and NLP solvers. The package PortfolioAnalytics by (Peterson and Carl 2015) uses standard linear and quadratic optimization tools (Rglpk and quadprog) and a number of new packages: stochastic optimization (DEoptim, GenSA) and particle swarm optimization (psoptim). The package parma by (Ghalanos 2016) offers scenario and moment based optimization of portfolios for a large class of risk and deviation measures using Rglpk, quadprog and, for non-linear optimization nloptr. The above mentioned packages are very effective in solving medium-size portfolio problems, however, due to their use of standard optimization tools they cannot deal with large problems either in the number of assets or the size of the distribution. Moreover, none of these packages is able to select an optimal portfolio which is closest to a given benchmark portfolio.
The new package
PortfolioOptim
overcomes the aforementioned limitations solving portfolio problems in
mean-risk setting with linear portfolio constraints and risk measures
that make the problem reducible to a linear program. Attempts to apply
LP solvers to more general portfolio problems (cf. (Mansini et al. 2014) and references
therein) are not included in the package. Our first contribution is an
efficient search algorithm for optimal portfolios in stochastic programs
with a very large number of scenarios. A large number of scenarios in
portfolio optimization appears often when a continuous distribution of
returns is approximated by a discrete one. The goal is to obtain an
optimal portfolio which approximates the optimal portfolio for the
continuous distribution. This can be achieved by performing optimization
on a large discrete sample generated from the continuous distribution
under consideration. A large sample leads to a high dimensional LP
problem which can be difficult to solve by general-purpose LP solvers.
This problem has been addressed by Künzi-Bay and Mayer (2006) who solved the portfolio
optimization problem in mean-CVaR setting using Benders decomposition.
They have computed accurately CVaR of the optimal portfolio using
samples of order
Our second contribution relates to portfolio optimization problems when only a small number of random scenarios is available. Small discrete samples in portfolio optimization appear when the distribution of returns is an empirical distribution of historical returns. Usually one takes 5 or 10 years of weekly or monthly returns giving between 120 and 500 samples. Solutions to linear programs of such dimensions are in many cases non-unique and can occupy a whole face of a multidimensional simplex. In such cases, a standard software finds only one solution, usually an vertex of the solution simplex. This is often not the optimal portfolio which asset managers consider as the most appropriate. Asset managers have usually certain beliefs about the composition of optimal portfolios and those can be expressed as similarity or vicinity of the desired optimal portfolio to some benchmark portfolio. Therefore, one wants to find an optimal portfolio which has the smallest distance to a given benchmark, or, equivalently, the orthogonal projection of the benchmark on the space of optimal portfolios.
In the LP literature a special case of the above projection problem has
been discussed for a long time. This is the problem of the least norm LP
solution, also called normal solution (cf. Zhao and Li (2002) and references cited
therein). Unfortunately, the algorithms which find normal solutions
cannot be easily adapted to our problem. The reason is that portfolio
weights make up only a fraction of all coordinates of the LP solution
(see the examples in Sections LP computable portfolio
problems and Benders decomposition). Contrary
to the normal solution which is the projection of the origin onto the
simplex of optimal solutions, we are looking for a projection onto a
subspace of portfolio weights. In addition, we are not looking for a
solution with the least norm but for a solution with the smallest
distance to an arbitrary vector (benchmark portfolio). It appears
however, that the regularized central path algorithm due to Zhao and Li (2002)
which solves simultaneously primal and dual LP problems finding the
least norm solutions to both problems, can be modified to our purposes.
Our contribution is the extension of this algorithm to the following
problem. Given the set
The rest of the paper is organized as follows. In Section LP computable portfolio problems, we describe the portfolio optimization problems that can be reduced to LP problems and are considered as working examples in the rest of the paper. These portfolio optimization problems are also implemented in the package PortfolioOptim. Section Benders decomposition analyzes Benders decomposition algorithm applied to simple recourse problems. As an illustration, we present applications to the portfolio optimization in mean-CVaR and mean-LSAD settings. Section Projection algorithm for portfolio optimization describes the adaptation of the path-following algorithm of Zhao and Li (2002) to the construction of an optimal portfolio with the smallest distance to a benchmark portfolio. We present also the proof of convergence for the modified algorithm. Computational examples and the analysis of performance of both algorithms are presented in Section Numerical examples. The paper ends with short Summary.
We consider the portfolio optimization problem in mean-risk setting. The
risk is measured by a risk (or deviation) measure
We will assume that at optimum the constraint
It appears that for a number of risk measures the optimization problem
(1) formulated for a discrete distribution
Consider the risk measured by conditional value at risk (CVaR). For a
random outcome
Consider now mean-MAD and mean-LSAD optimization problems. MAD (mean
absolute deviation) is defined as (cf. Konno and Yamazaki (1991))
For LSAD (lower semi absolute deviation) as defined by Konno (1990) or Konno et al. (2002)
In this section, we present a solution to a simple linear recourse
problem with random technology matrix when the distribution of
stochastic variable is represented by a large number of scenarios. The
simple recourse problem is a special case of two-stage recourse problem
We see that each pair of recourse variables
For each optimization problem
For a discrete probability space with
This problem is even harder than the original linear program
(13) as there are
Let
The complete algorithm is as follows:
Initialization: set
Solve problem (16) and denote the solution by
If
Set
The algorithm written as an R script can be seen on Fig. 1.
We begin with an example of mean-CVaR optimization for a discrete
distribution of returns. Our goal is to find portfolio
Due to the relation
As we have discussed earlier, some mean-risk optimization problems can
be transformed into linear programs. Those linear programs can be
written down in the standard form
In practical computations, it appears that solutions to the above linear program are sometimes non-unique and form a (multidimensional) simplex. A typical linear solver find only one of the solutions – usually one of the vertices of the solution simplex. In this section our aim is to find a solution to the above linear program which is closest to a given vector. It often lies in the interior of a face of the solution simplex.
Let
The least norm solution to linear program mentioned in the Introduction
is a special case of problem
((21)–(22)). It corresponds to
To solve problem ((21)–(22)) we follow
the approach by Zhao and Li (2002) based on the path-following algorithm. We begin
with the reformulation of (21) as the logarithmic
barrier problem
The Lagrangian for this problem reads
To find the solution to (23) which fulfills the
condition
To simplify the problem we assume that
The complete algorithm is as follows:
Initialization. Set
Newton’s iterates. If
Otherwise, find
Then set
Reduction of
We prove now that starting from the initial iterate defined in Step 1,
we obtain a convergent sequence of iterates. The estimate of the norm
Lemma 1. The central path projection algorithm is well-defined. The
sequence
Proof. The proof goes along the lines of the original proof of
(Zhao and Li 2002). We have only to remark that the nonsingularity of matrix
For the correctness of Step 3 of the algorithm, we have to show that
point
The proof of the convergence of the iterative sequence requires some modification of the proof by Zhao and Li (2002). The modification is formulated in the following lemma.
Lemma 2. When the solution set to equation (27) is
nonempty, then the sequence
Proof. We follow the line of the proof of Theorem 4.1 in (Zhao and Li 2002).
Let
Let
In what follows, we use the positive semidefiniteness
Finally we obtain
The boundedness of
The complete algorithm is presented on Fig. 2. This
algorithm requires an appropriate initial step:
The choice of
Set
Set
Take
Set
Compute
Set
Set
Compute
Set
Remark 3. Due to the above choice of the initial point we have the
estimates
The computational complexity of our algorithm arises from the solution
of the
For
We show now how the general projection algorithm described above can be
used to find an optimal portfolio with the smallest distance to a given
benchmark portfolio. We begin with the mean-CVaR optimization of Section
LP computable portfolio problems. For simplicity, we take
A solution to this problem can be obtained by the central path
projection algorithm. To this end, we define
With the above definitions, the solution obtained by the central path projection algorithm is an optimal solution to problem (32) with constraint (33).
For the mean-LSAD problem with
The algorithms described in Sections Benders decomposition and Projection algorithm for portfolio optimization are implemented in R in the package PortfolioOptim. The package offers two public functions:
BDportfolio_optim
– which performs portfolio optimization using
Benders decomposition;
PortfolioOptimProjection
– which implements the projection
algorithm described in Section Projection algorithm for portfolio
optimization.
For solving LP subproblems in BDportfolio_optim
, we use function
Rglpk_solve_LP
from the R-package Rglpk. All computations are
carried out on a computer with Intel Core i5-5200U processor with 8 GB
RAM running Linux operating system.
It has been already observed by Künzi-Bay and Mayer (2006) that a good discrete approximation
of a continuous distribution of returns requires large samples. The size
of the sample depends on the goal of optimization procedure. It has been
shown in Künzi-Bay and Mayer (2006) that the value of the objective function is accurately
estimated with samples of size
Consider first the model used in Künzi-Bay and Mayer (2006): they provide the vector of mean
First we initialize computations with Künzi-Bay and Mayer (2006) data.
library(mvtnorm)
library(PortfolioOptim)
generate\_data\_normal <- function (means, covmat, num)
\{
k <- ncol(covmat)
sim\_data <- rmvnorm (n=num, mean = means, sigma=covmat)
sim\_data <- matrix(num, k, data = sim\_data)
colnames(sim_data) <- colnames(covmat)
prob <- matrix(1/num,num,1)
mod\_returns <- cbind(sim\_data, prob)
return (mod\_returns)
\}
prepare\_data\_KM <- function ()
\{
sample\_cov <- matrix(5,5, data = c( 0.003059 , 0.002556 , 0.002327 , 0.000095 , 0.000533,
0.002556 , 0.003384 , 0.002929 , 0.000032 , 0.000762,
0.002327 , 0.002929 , 0.003509 , 0.000036 , 0.000908,
0.000095 , 0.000032 , 0.000036 , 0.000069 , 0.000048 ,
0.000533 , 0.000762 , 0.000908 , 0.000048 , 0.000564))
sample\_mean <- c( 0.007417, 0.005822, 0.004236, 0.004231, 0.005534)
colnames(sample\_cov) <- c("MSCI.CH", "MSCI.E", "MSCI.W", "Pictet.Bond", "JPM.Global")
return(list( sample\_mean = sample\_mean, sample\_cov = sample\_cov))
\}
We perform two tests (both tests are run simultaneously). In the first test, we compare computational time for different sample sizes. In the second test, we compare the accuracy of the obtained optimal portfolios. We perform computations for different sample sizes; for each sample size we generate 10 independent samples and assess the mean and the variance of the running time and portfolio weights. The R code for these tests is as follows.
data\_nA <- prepare\_data\_KM()
sample\_cov <- data\_nA\$sample\_cov
sample\_mean <- data\_nA\$sample\_mean
k <- ncol(sample\_cov)
a0 <- rep(1,k)
Aconstr <- rbind(a0,-a0)
bconstr <- c(1+1e-8, -1+1e-8)
lbound <- rep(0,k)
ubound <- rep(1,k)
R0 = 0.005
ET <- NULL
weights <- NULL
repetition = 10
sample\_size = 10 000 \# also run for 100 000 and 1 000 0000
ptm <- proc.time()[3]
for (i in 1:repetition )\{
mod\_returns <- generate\_data\_normal (sample\_mean, sample\_cov, sample\_size)
res <- BDportfolio\_optim (mod\_returns, R0, risk = "CVAR", alpha = 0.95,
Aconstr, bconstr,lbound, ubound, maxiter = 200, tol = 1e-10 )
ET <- c(ET, proc.time()[3] - ptm)
ptm <- proc.time()[3]
weights <- rbind(weights, t(res\$theta))
\}
cat(''running time and its standard deviation \(\backslash\)n'')
print(mean(ET))
print(sqrt(var(ET)))
cat(''optimal portfolio and confidence intervals of its weights \(\backslash\)n'')
print((colMeans(weights))*100)
print(sqrt(apply(weights, 2, var))*100*4/sqrt(repetition))
Since standard LP solvers (used for comparison by Künzi-Bay and Mayer (2006)) are too memory demanding to run on our hardware, we report computational time only for our BDportfolio_optim function. Table 1 contains the average running times for 10 different realizations together with the standard deviations.
sample size | mean | st. dev. |
10 000 | 0.0663 | 0.0085 |
100 000 | 0.4953 | 0.0363 |
1 000 000 | 4.518 | 0.1086 |
The average running time for samples of size
The estimates of optimal portfolio weights for different sample sizes
are collected in Table 2. The 95% confidence intervals of
portfolio weights are reported in brackets. If the difference between
portfolio weights is larger that two standard deviations we conclude
that the difference is statistically significant (this corresponds to
95% two-sided Student-t test whose critical value for the sample of size
10 is 2.228). The results of Table 2 show that samples of size
sample size | |||
assets | |||
MSCI.CH | 9.6 | 11.2 | 10.9 |
(4.33) | (1.45) | (0.39) | |
MSCI.E | 0.0 | 0.0 | 0.0 |
(0.00) | (0.00) | (0.00) | |
MSCI.W | 0.0 | 0.0 | 0.0 |
(0.00) | (0.00) | (0.00) | |
Pictet.Bond | 53.6 | 56.2 | 56.8 |
(13.24) | (2.33) | (0.83) | |
JPM.Global | 36.8 | 32.6 | 32.3 |
(11.41) | (2.37) | (0.74) |
We analyze now the effect of the number of assets in portfolio on the
accuracy of computations. To this end, we use the data-set etfdata
from the R-package parma. For comparison we take two subsets:
etfdata[1:500,1:5]
(with 5 assets) and etfdata[1:500,1:10]
(with 10
assets). We add to our code a new function computing the vector of means
and covariance matrix for a data set with
library(parma)
library(xts)
library(quantmod)
data(etfdata)
prepare\_data <- function (k)
\{
quotData <- as.xts(etfdata[1:500, 1:k])
retData = NULL
for (i in 1:k){
retData <- cbind(retData,weeklyReturn(quotData[,i], type='arithmetic'))
}
colnames(retData) <- colnames(quotData)
sample_cov <- cov(retData)
sample_mean <- colMeans(retData)
return(list(sample\_mean = sample\_mean, sample\_cov = sample\_cov))
\}
data\_nA <- prepare\_data(5) \# also run with prepare\_data(10)
sample\_cov <- data\_nA\$sample\_cov
sample\_mean <- data\_nA\$sample\_mean
k <- ncol(sample\_cov)
a0 <- rep(1,k)
Aconstr <- rbind(a0,-a0)
bconstr <- c(1+1e-8, -1+1e-8)
lbound <- rep(0,k)
ubound <- rep(1,k)
R0 = 0.004
sample\_size = 10 000 \# also run for 100 000 and 1 000 0000
repetition = 100
weights <- NULL
for (i in 1:repetition )\{
returns <- generate\_data\_normal(sample\_mean, sample\_cov, sample\_size)
res <- BDportfolio_optim (returns, R0, risk = "CVAR", alpha = 0.95,
Aconstr, bconstr, lbound, ubound, maxiter = 200, tol = 1e-10 )
weights <- rbind(weights, t(res\$theta))
\}
print(sum(sqrt(apply(weights, 2, var))*100*4/sqrt(repetition))/k)
For each data set we generate samples of size
sample size | |||
assets number | |||
5 | 0.97 | 0.29 | 0.09 |
10 | 1.18 | 0.42 | 0.15 |
In testing the projection algorithm described in Section Projection
algorithm for portfolio optimization the following
values of parameters are used etfdata[1:500,1:5]
and etfdata[1:500,1:10]
from the package
parma. Samples of 125, 250 and 500 returns are generated assuming
they are weekly returns (the vector of means and covariance matrix
appropriately scaled) which corresponds to 2.5, 5 and 10 years of weekly
data. For each sample we compute the vector of portfolio weights using
the function PortfolioOptimProjection with tol = 1e-6 and taking as the
benchmark the portfolio
The code is as follows
data\_nA <- prepare\_data(5) \# also run with prepare\_data(10)
sample\_cov <- data\_nA\$sample\_cov
sample\_mean <- data\_nA\$sample\_mean
k <- ncol(sample\_cov)
a0 <- rep(1,k)
Aconstr <- rbind(a0,-a0)
bconstr <- c(1+1e-8, -1+1e-8)
lbound <- rep(0,k)
ubound <- rep(1,k)
w\_m = rep(1/k,k)
R0 = 0.005
returns_size =125 \# also run for 250 and 500
ET <- NULL
ptm <- proc.time()[3]
for (i in 1:10 )\{
mod\_returns <- generate\_data\_normal(sample\_mean, sample\_cov, returns_size )
res <- PortfolioOptimProjection (mod\_returns, R0 , risk = "CVAR", alpha = 0.95,
w\_m, Aconstr, bconstr,lbound, ubound, 800, tol = 1e-6 )
ET <- c(ET, proc.time()[3] - ptm)
ptm <- proc.time()[3]
\}
cat(''Mean running time and its standard deviation \(\backslash\)n'')
print(c( mean(ET), sqrt(var(ET))))
The average running time per one computation and its standard deviation is reported in Table 4.
number of assets | sample size | mean | sd. deviation |
5 | 125 | 6.8618 | 5.7377 |
250 | 35.1910 | 30.2499 | |
500 | 134.9312 | 95.2407 | |
10 | 125 | 3.1451 | 1.6976 |
250 | 30.0261 | 22.5226 | |
500 | 125.1935 | 104.7991 |
Taking into account that the sample of size
We presented two optimization algorithms for financial portfolios. These algorithms find optimal portfolios in problems when the nonlinear optimization in mean-risk setting can be reformulated as a linear programming problem. The algorithm implementing Benders cuts allows for large samples. Our experience with very large data samples obtained by simulation from a given distribution shows that the algorithm is robust and estimated optimal portfolios are very stable (the standard deviation of portfolio weights is below 0.5 %). The second algorithm based on the central path projection can be useful for small data samples where solutions are not unique and form a multidimensional simplex. Extracting from this simplex a point with the smallest distance to a given benchmark portfolio can be used to improve the decision process in asset management. However, the second algorithm which implements the central path projection requires further analysis. In particular, we would like to make the algorithm’s performance and running time less dependent on the realization of the data.
The author gratefully acknowledges financial support from National Science Centre, Poland, project 2014/13/B/HS4/00176.
fPortfolio, PortfolioAnalytics, Rglpk, quadprog, DEoptim, GenSA, psoptim, parma, nloptr, PortfolioOptim
This article is converted from a Legacy LaTeX article using the texor package. The pdf version is the official version. To report a problem with the html, refer to CONTRIBUTE on the R Journal homepage.
Text and figures are licensed under Creative Commons Attribution CC BY 4.0. The figures that have been reused from other sources don't fall under this license and can be recognized by a note in their caption: "Figure from ...".
For attribution, please cite this work as
Palczewski, "LP Algorithms for Portfolio Optimization: The PortfolioOptim Package", The R Journal, 2018
BibTeX citation
@article{RJ-2018-028, author = {Palczewski, Andrzej}, title = {LP Algorithms for Portfolio Optimization: The PortfolioOptim Package}, journal = {The R Journal}, year = {2018}, note = {https://rjournal.github.io/}, volume = {10}, issue = {1}, issn = {2073-4859}, pages = {308-327} }