LP Algorithms for Portfolio Optimization: The PortfolioOptim Package

Abstract:

The paper describes two algorithms for financial portfolio optimization with the following risk measures: CVaR, MAD, LSAD and dispersion CVaR. These algorithms can be applied to discrete distributions of asset returns since then the optimization problems can be reduced to linear programs. The first algorithm solves a simple recourse problem as described by Haneveld using Benders decomposition method. The second algorithm finds an optimal portfolio with the smallest distance to a given benchmark portfolio and is an adaptation of the least norm solution (called also normal solution) of linear programs due to Zhao and Li. The algorithms are implemented in R in the package PortfolioOptim.

Cite PDF Tweet

Published

May 21, 2018

Received

Jul 23, 2017

Citation

Palczewski, 2018

Volume

Pages

10/1

308 - 327


1 Introduction

The construction of an optimal portfolio of financial instruments is one of primal goals in asset management. Recent advances in risk measurement advocate using risk measures that take into account the tail behavior of asset return distributions, such as conditional value-at-risk or lower semi-absolute deviation. For these risk measures, finding optimal portfolios in risk-return setting leads to convex programming problems. For a finite discrete distribution of returns and many practically used risk measures the optimization problem can be solved by LP methods. The portfolio optimization by LP methods leads to two problems: for distributions with a large number of values (for example, a large sample from a continuous distribution) one obtains a unique solution but the LP problem is resource demanding; for distributions with a small number of values one often obtains a non-unique solution, which is of limited use for asset management. In the latter case, a unique optimal portfolio is obtained by projecting a benchmark portfolio on the whole set of solutions. In the paper we present an algorithm that solves efficiently the problem with large distributions and an algorithm that finds an orthogonal projection of the benchmark portfolio on the space of solutions.

The R language and environment for statistical computing offer a large variety of tools for portfolio optimization. General purpose optimization tools are reviewed by (R packages for solving optimization problems) and (R packages for convex optimization). The book by provides an overview of specific functions for portfolio optimization which are embedded in financial packages. Here we mention a selection of R-packages dedicated primarily to portfolio optimization. The package fPortfolio by offers a large set of functions for financial data analysis and enables portfolio optimization in mean-variance, mean-MAD and mean-CVaR settings. For these portfolio problems the package employs existing optimization tools: LP, QP and NLP solvers. The package PortfolioAnalytics by uses standard linear and quadratic optimization tools (Rglpk and quadprog) and a number of new packages: stochastic optimization (DEoptim, GenSA) and particle swarm optimization (psoptim). The package parma by offers scenario and moment based optimization of portfolios for a large class of risk and deviation measures using Rglpk, quadprog and, for non-linear optimization nloptr. The above mentioned packages are very effective in solving medium-size portfolio problems, however, due to their use of standard optimization tools they cannot deal with large problems either in the number of assets or the size of the distribution. Moreover, none of these packages is able to select an optimal portfolio which is closest to a given benchmark portfolio.

The new package PortfolioOptim overcomes the aforementioned limitations solving portfolio problems in mean-risk setting with linear portfolio constraints and risk measures that make the problem reducible to a linear program. Attempts to apply LP solvers to more general portfolio problems (cf.  and references therein) are not included in the package. Our first contribution is an efficient search algorithm for optimal portfolios in stochastic programs with a very large number of scenarios. A large number of scenarios in portfolio optimization appears often when a continuous distribution of returns is approximated by a discrete one. The goal is to obtain an optimal portfolio which approximates the optimal portfolio for the continuous distribution. This can be achieved by performing optimization on a large discrete sample generated from the continuous distribution under consideration. A large sample leads to a high dimensional LP problem which can be difficult to solve by general-purpose LP solvers. This problem has been addressed by who solved the portfolio optimization problem in mean-CVaR setting using Benders decomposition. They have computed accurately CVaR of the optimal portfolio using samples of order 104. Our extension of this result is twofold. First, we design an algorithm that implements Benders decomposition for a general class of simple recourse problems. Portfolio optimization problems for the risk measures mentioned above are special cases of these simple recourse problems. Second, using an internal point LP solver from GLPK library and appropriately modifying the stopping criterion we substantially increase the size of discrete samples which can be used in computations. Our algorithm can perform computations for samples of order 106 on a standard computer in a few seconds.

Our second contribution relates to portfolio optimization problems when only a small number of random scenarios is available. Small discrete samples in portfolio optimization appear when the distribution of returns is an empirical distribution of historical returns. Usually one takes 5 or 10 years of weekly or monthly returns giving between 120 and 500 samples. Solutions to linear programs of such dimensions are in many cases non-unique and can occupy a whole face of a multidimensional simplex. In such cases, a standard software finds only one solution, usually an vertex of the solution simplex. This is often not the optimal portfolio which asset managers consider as the most appropriate. Asset managers have usually certain beliefs about the composition of optimal portfolios and those can be expressed as similarity or vicinity of the desired optimal portfolio to some benchmark portfolio. Therefore, one wants to find an optimal portfolio which has the smallest distance to a given benchmark, or, equivalently, the orthogonal projection of the benchmark on the space of optimal portfolios.

In the LP literature a special case of the above projection problem has been discussed for a long time. This is the problem of the least norm LP solution, also called normal solution (cf.  and references cited therein). Unfortunately, the algorithms which find normal solutions cannot be easily adapted to our problem. The reason is that portfolio weights make up only a fraction of all coordinates of the LP solution (see the examples in Sections LP computable portfolio problems and Benders decomposition). Contrary to the normal solution which is the projection of the origin onto the simplex of optimal solutions, we are looking for a projection onto a subspace of portfolio weights. In addition, we are not looking for a solution with the least norm but for a solution with the smallest distance to an arbitrary vector (benchmark portfolio). It appears however, that the regularized central path algorithm due to which solves simultaneously primal and dual LP problems finding the least norm solutions to both problems, can be modified to our purposes. Our contribution is the extension of this algorithm to the following problem. Given the set S of optimal solutions to an LP problem, find xS such that B(xx^)B(xx^) for all xS, where x^ is a given vector and B is the operator of projection on a subspace. This general algorithm is then adapted to obtain an optimal portfolio with the smallest distance to a benchmark (see Section Projection algorithm for portfolio optimization for details).

The rest of the paper is organized as follows. In Section LP computable portfolio problems, we describe the portfolio optimization problems that can be reduced to LP problems and are considered as working examples in the rest of the paper. These portfolio optimization problems are also implemented in the package PortfolioOptim. Section Benders decomposition analyzes Benders decomposition algorithm applied to simple recourse problems. As an illustration, we present applications to the portfolio optimization in mean-CVaR and mean-LSAD settings. Section Projection algorithm for portfolio optimization describes the adaptation of the path-following algorithm of to the construction of an optimal portfolio with the smallest distance to a benchmark portfolio. We present also the proof of convergence for the modified algorithm. Computational examples and the analysis of performance of both algorithms are presented in Section Numerical examples. The paper ends with short Summary.

2 LP computable portfolio problems

We consider the portfolio optimization problem in mean-risk setting. The risk is measured by a risk (or deviation) measure R (for the definitions of risk and deviation measures the reader is advised to consult the paper by ). Let R be a distribution of returns of J risky assets. We denote by R^ centered returns, i.e. R^=RE[R]. Consider the problem of finding an optimal portfolio u which fulfills the conditions (1){R(uTR)min,uTE[R]r0,uX, where X is a polyhedral set (i.e. given by linear constraints) and r0 is the required return of the portfolio. The set X is usually the whole space (when no limitations on trading positions are imposed) or the positive orthant (when short-selling of assets is prohibited).

We will assume that at optimum the constraint uTE[R]r0 is binding. This is the case when r0 is not smaller than the return for the minimal risk portfolio, i.e. the portfolio which solves the problem minuXR(uTR).

It appears that for a number of risk measures the optimization problem (1) formulated for a discrete distribution R can be reduced to a linear program. Let the asset returns R be given by a discrete distribution placing weights pn at points rn, n=1,,N. When the distribution is centered points of R^ will be denoted by r^n.

Consider the risk measured by conditional value at risk (CVaR). For a random outcome X we define VaRα(X) as VaRα(X)=inf{z: P(Xz)>α}. Then the conditional value at risk is defined as CVaRα(X)=1α0αVaRp(X)dp. observed that for X=uTR the minimization of CVaR can be formulated as the following nonlinear problem (2)minu,ξξ+11αE[(uTRξ)+], where the latent variable ξ corresponds to VaRα(uTR). For discrete distributions and the risk measured by CVaR, the portfolio optimization problem can therefore be reformulated as the linear program (3){ξ+11αn=1Npnynmin,yn0,n=1,,N,yn+ξ+rnTu0,n=1,,N,μTur0,uX. Here μ=E[R], while yn plays a role of the shortfall of return uTrn below ξ: yn=(ξuTrn)+=(ξ+uTrn). For deviation CVaR we replace rn by r^n in (3).

Consider now mean-MAD and mean-LSAD optimization problems. MAD (mean absolute deviation) is defined as (cf. ) MAD(uTR)=E[|uTRE[uTR]|]=E[|uTR^|]. For discrete distributions mean-MAD optimization problem reads (4){n=1Npn|r^nTu|min,μTur0,uX. This problem can be formulated as the linear program (5){n=1Npnynmin,yn0,n=1,,Nyn+r^nTu0,n=1,,N,ynr^nTu0,n=1,,N,μTur0,uX.

For LSAD (lower semi absolute deviation) as defined by or LSAD(uTR)=E[|uTRE[uTR]|]=E[|uTR^|], mean-LSAD optimization has the form (6){n=1Npn|r^nTu|min,μTur0,uX. This leads to the following linear program (7){n=1Npnynmin,yn0,n=1,,N,yn+r^nTu0,n=1,,N,μTur0,uX. Since LSAD(uTR)=12MAD(uTR), in what follows we shall consider only one of the above optimization problems.

3 Benders decomposition

Algorithm

In this section, we present a solution to a simple linear recourse problem with random technology matrix when the distribution of stochastic variable is represented by a large number of scenarios. The simple recourse problem is a special case of two-stage recourse problem (8){cTx+Q(x)min,xX, with the following form of the recourse subproblem (9){Q(x)=E[v(x)],v(x)=miny{(q+)Ty++(q)Ty:y+y=bAx, y+,yR+m}. In this framework, q+ and q are known penalty costs, matrix A=A(ω) and vector b=b(ω) are random and X is a bounded, convex subset in Rl.

We see that each pair of recourse variables (yi+,yi), i=1,,m , depends only on the i-th row in the condition bAx, so that their optimal values can be determined independently. Thus, the second-stage value function v(x) is separated into the sum of m functions v(x)=i=1mminyi{qi+yi++qiyi:yi+yi=biAix,yi+,yi0}, where Ai is the ith row of matrix A and bi is the ith element of vector b.

For each optimization problem (10)minyi{qi+yi++qiyi:yi+yi=biAix,yi+,yi0}, the dual problem has the form supλi{λi(biAix):λiqi+,λiqi}. The dual problem is feasible only if qi++qi0. When this condition holds, the solution to the dual problem is given by λi={qi+,if biAix>0,qi,if biAix0, and the optimal solution to problem (10) has the form yi+=max(0,(biAix)),yi=max(0,(biAix)). Then the value function Q(x) can be written as (11)Q(x)=i=1m(qi+E[(biAix)+]+qiE[(biAix)])=i=1m(qi+E[(Aixbi)]+qiE[(Aixbi)+]). Since s+s=s, we have (12)Q(x)=i=1m(qi(A¯ixb¯i)+(qi++qi)E[(Aixbi)])=(q)TA¯x(q)Tb¯+E[(q++q)T(bAx)+], where A¯i=E[Ai] and b¯i=E[bi].

For a discrete probability space with P(ω=ωn)=pn, n=1,,N, the simple recourse problem can be reformulated in the following way (13){cTx+(q)T(A¯xb¯)+n=1Npnhnmin,hn(q++q)T(b(ωn)A(ωn)x),hn0,xX. This linear program with a large size N of the probability space is difficult to solve due to the number of constraints. solved it using a special version of the L-shaped method. Their approach was further extended by who applied directly Benders decomposition (cf., ). In terms of Benders cuts (13) is equivalent to (we skip (q)Tb¯ which is constant) (14){minx,wcTx+(q)TA¯x+w,wnKpn(q++q)T(b(ωn)A(ωn)x),for all KM,w0,xX, where M={1,,N}.

This problem is even harder than the original linear program (13) as there are 2N constraints, but showed that the above problem can be solved through a sequence of expanding relaxed problems (15){minx,wcTx+(q)TA¯x+w,wnKkpn(q++q)T(b(ωn)A(ωn)x),k=1,,ν,w0,xX, ν is the number of steps, KkM are constraints added in step k with KkKl when kl. Hence, each successive relaxed problem adds more constraints to those already present in the previous steps.

Let ηk=nKkpn(q++q)TA(ωn),ζk=nKkpn(q++q)Tb(ωn). Then (15) is written as the linear program (16){minx,wcTx+(q)TA¯x+w,wζkηkx,k=1,,ν,w0,xX.

The complete algorithm is as follows:

  1. Initialization: set K1=M, η1=n=1Npn(q++q)TA(ωn), ζ1=n=1Npn(q++q)Tb(ωn) and ν=1. Choose computation accuracy ε.

  2. Solve problem (16) and denote the solution by (x,w). Put $$

    K={nM:(q++q)T(b(ωn)A(ωn)x)0},w+=nKpn(q++q)T(b(ωn)A(ωn)x),w^=ζνηνx.

    Compute = cTx+ (q-)T{A} x^ + w_+^, = cTx + (q-)T{A} x^* + w^*.$$

  3. If (FF)ε then stop. x is an optimal solution.

  4. Set ν=ν+1, Kν=K. Compute ην=nKνpn(q++q)TA(ωn) $$ and ζν=nKνpn(q++q)Tb(ωn). Add this new constraint to the set of constraints and go to Step 2.

The algorithm written as an R script can be seen on Fig. 1.

Figure 1: Benders decomposition algorithm for a simple recourse problem

Examples

Conditional value at risk (CVaR)

We begin with an example of mean-CVaR optimization for a discrete distribution of returns. Our goal is to find portfolio u which solves (17){CVaRα(uTR)min,uTE[R]r0,uX. As is shown in Section LP computable portfolio problems, this problem is reduced to the linear program {ξ+11αn=1Npnynmin,yn0,n=1,,N,yn+ξ+rnTu0,n=1,,N,uTμr0,uX, where μ=E[R]. Benders decomposition for the above problem leads to the sequence of expanding relaxed problems (18){minu,ξ,wξ+11αw,nKkpn(ξ+rnTu)+w0,k=1,,ν,uTμr0,uX. Using notation from Section Benders decomposition we identify x=(u,ξ), q=0, (q+)TA(ωn)=(rn,1), b(ωn)=0 and c=(0J,1), where 0J is the zero vector of length J. The algorithm can be readily applied.

Mean absolute deviation(MAD) and lower semi-absolute deviation (LSAD)

Due to the relation LSAD(uTR)=12MAD(uTR) it is sufficient to consider only the mean-LSAD portfolio optimization: (19){LSAD(uTR)min,uTμr0,uX. As before, for a discrete set of returns that problem can be formulated as a linear program (see equation (7)) and Benders decomposition can be applied leading to a sequence of expanding relaxed problems. Introducing the new variable ηk=nKkpnr^n, we reduce the problem to the linear program (20){minu,ww,ηkTu+w0,k=1,,ν,μTur0,uX. Similarly as for CVaR, we recognize in that formulation the Benders problem from Section Benders decomposition.

4 Projection algorithm for portfolio optimization

Algorithm

As we have discussed earlier, some mean-risk optimization problems can be transformed into linear programs. Those linear programs can be written down in the standard form (21)cTxmin,Axb,x0,xRn.

In practical computations, it appears that solutions to the above linear program are sometimes non-unique and form a (multidimensional) simplex. A typical linear solver find only one of the solutions – usually one of the vertices of the solution simplex. In this section our aim is to find a solution to the above linear program which is closest to a given vector. It often lies in the interior of a face of the solution simplex.

Let S denote the set of optimal solutions to (21), B be a given invertible matrix in Rn and x^ a vector from Rn. We want to find xS such that (22)B(xx^)B(xx^) for all xS.

The least norm solution to linear program mentioned in the Introduction is a special case of problem ((21)(22)). It corresponds to x^=0 and B being an identity matrix. Finding an optimal portfolio closest to a given benchmark in the framework of LP problems can also be reduced, after some modification, to the solution of problem ((21)(22)). To describe this modification observe that for LP computable portfolio problems the independent variable x in (21) contains more coordinates that only portfolio weights. Let x=(x,u) where u corresponds to portfolio weights and x contains all other coordinates (for CVaR optimization (3) x=(ξ,y,u), hence x=(ξ,y), for MAD optimization (5) and LSAD optimization (7) x=(y,u)). Given the benchmark portfolio u^, the objective is to find an optimal portfolio which minimizes the distance uu^. To achieve that goal by solving problem ((21)(22)) we have to extend u^ to a vector x^=(x^,u^) on the whole space Rn taking arbitrary x^ and projecting the difference (xx^) on the subspace spanned by the portfolio coordinates. Let B be this projection operator, then the aim is to minimize B(xx^). Matrix B is diagonal with entries 1 on the diagonal positions corresponding to portfolio weights and 0 otherwise. Since B is not invertible the considered portfolio problem cannot be reduced to the solution of problem ((21)(22)). Hence we introduce a matrix B replacing in B zero diagonal entries with some small positive number ε. Then B is invertible and an optimal portfolio close to a benchmark can be found by solving problem ((21)(22)). One can further improve the accuracy of the computation by redoing the optimization with x^=(x,u^), where x is the solution obtained in the first optimization.

To solve problem ((21)(22)) we follow the approach by based on the path-following algorithm. We begin with the reformulation of (21) as the logarithmic barrier problem (23)cTxρ(i=1nlogxi+i=1mlogzi)min,Axz=b,x,z>0,xRn,zRm, where z is introduced to replace the inequality Axb by equality and ρ>0 is a regularizing parameter.

The Lagrangian for this problem reads L(x,y,z)=cTx+yT(zAx+b)ρ(i=1nlogxi+i=1mlogzi) and the Kuhn-Tucker solvability conditions are (24)diag(x)s=ρe,diag(z)y=ρe,s+ATyc=0,zAx+b=0, where diag(u) denotes the diagonal matrix with vector u on diagonal; the new variable s=ρdiag(x)1e is introduced to obtain the canonical representation of the central path approach; and e=(1,,1) (of an appropriate dimension).

To find the solution to (23) which fulfills the condition B(xx^)2min for a given vector x^ and matrix B, we modify the Lagrangian LP(x,y,z)=cTx+yT(zAx+b)ρ(i=1nlogxi+i=1mlogzi)+12θ(B(xx^)2y2). The stationary point of this Lagrangian gives the primal solution for which B(xx^)2 achieves minimal value and the dual solution which has minimal norm. The Kuhn-Tucker conditions give (25)diag(x)s=ρe,diag(z)y=ρe,s+ATyc=θBTB(xx^),zAx+b=θy.

To simplify the problem we assume that θ=κρp for given constants κ(0,1] and p(0,1). We define the nonlinear mapping (26)Fρ(x,y,s,z)=(diag(x)sρediag(z)yρes+ATycκρpBTB(xx^)zAx+bκρpy). The problem is now to find (x,y,s,z)0 which solve the equation (27)F0(x,y,s,z)=0. The solution to this equation is searched by the Newton iterative method Fρk(xk,yk,sk,zk)+Fρk(xk,yk,sk,zk)(Δx,Δy,Δs,Δz)=0, where Fρ(x,y,s,z) denotes the gradient of function Fρ given by the expression Fρ(x,y,s,z)=(diag(s)0diag(x)00diag(z)0diag(y)κρpBTBAT10Aκρp101). These iterative solutions form the path-following algorithm in a neighborhood of the regularized central path Nβ(ρ)={(x,y,s,z):Fρ(x,y,s,z)βρ}, with β(0,1).

The complete algorithm is as follows:

Central path projection algorithm

  1. Initialization. Set β(0,1) and assign scalars b1, b2 and σ in (0,1). Select (x0,y0,s0,z0)>0, κ(0,1), p(0,1) and ρ0(1,) such that (x0,y0,s0,z0)Nβ(ρ0).

  2. Newton’s iterates. If Fρk(xk,yk,sk,zk)=0 put (xk+1,yk+1,sk+1,zk+1)=(xk,yk,sk,zk) and go to Step 3.
    Otherwise, find (Δx,Δy) which solve the following linear system (28)(diag(sk)+κ(ρk)pBTBdiag(xk)diag(xk)ATdiag(yk)Adiag(zk)+κ(ρk)pdiag(yk)) (ΔxΔy)=(ρkediag(xk)skρkediag(zk)yk)(cdiag(xk)(κ(ρk)pBTB(xkx^)ATyksk+c)diag(yk)(Axk+κ(ρk)pykzkb)).

    Then set (29)(ΔsΔz)=(κ(ρk)pBTBATAκ(ρk)p1) (ΔxΔy)+(κ(ρk)pBTB(xkx^)ATyksk+cAxk+κ(ρk)pykzkb). Find α such that (xk+λΔx,yk+λΔy,sk+λΔs,zk+λΔz)>0 for all λ(0,α). Then find the maximal improvement step for λ by the Armijo rule: find the smallest j such that λj=αb1j and (30)Fρk(xk+λjΔx,yk+λjΔy,sk+λjΔs,zk+λjΔz)(1σλj)Fρk(xk,yk,sk,zk) Set (xk+1,yk+1,sk+1,zk+1)=(xk,yk,sk,zk)+λj(Δx,Δy,Δs,Δz) and go to Step 3.

  3. Reduction of ρ. Find the maximal improvement step for ρk by the Armijo rule: find the smallest j such that γj=b2j and F(1γj)ρk(xk+1,yk+1,sk+1,zk+1)β(1γj)ρk. Set ρk+1=(1γj)ρk and go to Step 2.

We prove now that starting from the initial iterate defined in Step 1, we obtain a convergent sequence of iterates. The estimate of the norm F creates a difficulty because of the term ρkp which for ρk<1 cannot be estimated by the first power of ρk. That requires such a choice of the starting point which cancels the terms with ρp. This goal is achieved by a proper choice of constant κ. The following lemma is an adaptation of Lemma 4.1 from .

Lemma 1. The central path projection algorithm is well-defined. The sequence ρk is monotonically decreasing and (xk,yk,sk,zk)Nβ(ρk) for all k0.

Proof. The proof goes along the lines of the original proof of . We have only to remark that the nonsingularity of matrix Fρ follows from the fact that the matrix (κρpBTBATAκρp1) is positive semidefinite for ρ>0.

For the correctness of Step 3 of the algorithm, we have to show that point (xk+1,yk+1,sk+1,zk+1) belongs to Nβ(ρk+1). To this end, we have to prove the estimate Fρk(x,y,s,z)Fρl(x,y,s,z)(ρlρk)+((ρl)p(ρk)p)(xx^,y), for ρlρk and (x,y,s,z)>0. Taking into account that κ1 that estimate follows from the definition of F Fρk(x,y,s,z)Fρl(x,y,s,z)(ρlρk)+κ((ρl)p(ρk)p)y+κ((ρl)p(ρk)p)BTB(xx^)(ρlρk)+((ρl)p(ρk)p)(xx^,y). The rest of the proof is similar as in . The boundedness of (xk,yk) follows from Lemma 2 below. ◻

The proof of the convergence of the iterative sequence requires some modification of the proof by . The modification is formulated in the following lemma.

Lemma 2. When the solution set to equation (27) is nonempty, then the sequence (xk,yk,sk,zk) obtained by the central path projection algorithm is bounded.

Proof. We follow the line of the proof of Theorem 4.1 in . Let (uk,vk,wk,qk) be defined as (uk,vk,wk,qk)=1ρkFρk(xk,yk,sk,zk). Then (uk,vk,wk,qk)β by the definition of Nβ(ρk) and the following system of equations holds (31)diag(xk)sk=ρk(e+uk),diag(yk)zk=ρk(e+vk),sk=ATyk+c+κ(ρk)pBTB(xkx^)+ρkwk,zk=Axkb+κ(ρk)pyk+ρkqk.

Let (x,y,s,z) be an optimal solution, i.e. a solution to equation (27). Then (x,y,s,z)0 and (s,z) is given by the expression (sz)=(0ATA0) (xy)+(cb). Due to (25) (x)Ts=0 and (y)Tz=0.

In what follows, we use the positive semidefiniteness (xy)T (κρpBTBATAκρp1) (xy)=κρpBx22+κρpy22. Then we have 0(xy)T (skzk)+(sz)T (xkyk)=(xxkyyk)T ((κ(ρk)pBTBATAκ(ρk)p1) (xkyk)+(cb)+ρk(wkqk)(κ(ρk)pBTBx^0))+(sz)T (xkyk)+(xkyk)T (skzk)=(xkxyky)T (κ(ρk)pBTBATAκ(ρk)p1) (xkxyky)(xkxyky)T((κ(ρk)pBTBATAκ(ρk)p1) (xy)+(cb)+ρk(wkqk)(κ(ρk)pBTBx^0))+(xkyk)T(sz)+(xkyk)T (skzk)=κ(ρk)pB(xkx)yky22κ(ρk)p(xkxyky)T (BTBx+(ρk)1pκ1wkBTBx^y+(ρk)1pκ1qk)+(xkyk)T (skzk). From equation (31) we obtain (xk)Tsk=ρkeT(e+uk),(yk)Tzk=ρkeT(e+vk) and the estimate |(xkyk)T (skzk)|ρkc for some positive constant c.

Finally we obtain B(xkx)yky22xkxyky2BTB(xx^)+(ρk)1pκ1wky+(ρk)1pκ1qk2+(ρk)1pκ1c. Since matrix B is invertible, we have x2B1Bx2. Taking into account the above estimates and the estimate ρkρ0, we obtain xkxyky2c.

The boundedness of (sk,zk) follows from the above estimate and equation (31). ◻

Figure 2: Regularized central path algorithm with projection

The complete algorithm is presented on Fig. 2. This algorithm requires an appropriate initial step: ρ0 and a starting point (x0,y0,s0,z0)Nβ(ρ0). Hence we choose (x0,y0,s0,z0)>0 such that Fρ0(x0,y0,s0,z0)βρ0, which guarantees that (x0,y0,s0,z0)Nβ(ρ0).

The choice of (x0,y0,s0,z0) starts with y0=e and z0=κρ0pe. We choose x0 and s0 so that the term κρ0pBTB(x0x^) is canceled. Taking x0=max(e,e+x^) makes x0>0. On the other hand, κρ0pBTB(x0x^) is a vector with nonzero components bounded from above by κρ0p(1+x^). Hence we define s0,init=BTB(x0x^) and compute κ=1/diag(x0)s0,init. Then we take s0=κρ0ps0,init. That procedure eliminates from Fρ0 the terms with ρ0p and replaces them by constants. We can now describe the construction of the initial point step-by-step:

  1. Set y0=e.

  2. Set x0=max(e,e+x^).

  3. Take ρ0=max(1,cATy0cAx0+b)+δ

  4. Set s0,init=BTB(x0x^).

  5. Compute κ=1/diag(x0)s0,init.

  6. Set z0=κρ0pe.

  7. Set s0=κρ0ps0,init.

  8. Compute ξ=Fρ0(x0,y0,s0,z0)ρ0.

  9. Set β(ξ,1).

Remark 3. Due to the above choice of the initial point we have the estimates diag(x0)s0ρ0e|ρ0ρ0p|ρ0,diag(z0)y0ρ0eκ|ρ0ρ0p|+(1κ)ρ0ρ0,s0+ATy0cκρ0pBTB(x0x^)ATy0c,z0Ax0+bκρ0py0Ax0+b. Hence Fρ0(x0,y0,s0,z0)ρ0 and for a properly chosen δ>0 the above inequality is sharp and ξ<1. Then taking β(ξ,1) we obtain the desired estimate Fρ0(x0,y0,s0,z0)βρ0.

The computational complexity of our algorithm arises from the solution of the (n+m)-dimensional system (28). We can reduce the dimension of this system observing that it can be written as (diag(sk)+κ(ρk)pBTBdiag(xk))Δxdiag(xk)ATΔy(diag(sk)+=ρkediag(xk)(κ(ρk)pBTB(xkx^)ATyk+c),diag(yk)AΔx+(diag(zk)+κ(ρk)pdiag(yk))Δy(diag(sk)+=ρkediag(yk)(Axk+κ(ρk)pykb). When nm eliminating Δx gives the equation HkΔy=ρkediag(yk)(Axk+κ(ρk)pykb)diag(yk)A×(diag(sk)+κ(ρk)pdiag(xk)BTB)1×(ρkediag(xk)(κ(ρk)pBTB(xkx^)ATyk+c)), where Hk=diag(zk)+κ(ρk)pdiag(yk)+diag(yk)A(diag(sk)+κ(ρk)pdiag(xk)BTB)1diag(xk)AT. For Δx we obtain then Δx=(diag(sk)+κ(ρk)pdiag(xk)BTB)1×(diag(xk)ATΔy+ρkediag(xk)(κ(ρk)pBTB(xkx^)ATyk+c)).

For m>n we can eliminate Δy to obtain MkΔx=ρkediag(xk)(κ(ρk)pBTB(xkx^)ATyk+c)+diag(xk)AT×(diag(zk)+κ(ρk)pdiag(yk))1(ρkediag(yk)(Axk+κ(ρk)pykb)), where Mk=diag(sk)+κ(ρk)pBTBdiag(xk)+diag(xk)AT(diag(zk)+κ(ρk)pdiag(yk))1diag(yk)A. For Δy we obtain then Δy=(diag(zk)+κ(ρk)pdiag(yk))1×(diag(yk)AΔx+ρkediag(yk)(Axk+κ(ρk)pykb)).

Examples

We show now how the general projection algorithm described above can be used to find an optimal portfolio with the smallest distance to a given benchmark portfolio. We begin with the mean-CVaR optimization of Section LP computable portfolio problems. For simplicity, we take X={u:u0}. Then the linear program corresponding to this problem reads (32){ξ+11αn=1Npnynmin,yn0,n=1,,N,yn+ξ+rnTu0,n=1,,N,uTμr0,u0. Given the benchmark portfolio wb, we are looking for a solution of the above linear program with the additional constraint (33)uwb2min.

A solution to this problem can be obtained by the central path projection algorithm. To this end, we define x=(u,ξ,y), c=(0J,1,p), where 0J is the zero vector of length J and p is the N dimensional vector with entries pn. Matrix A and vector b are given by the expressions A=(r0NTdiag(eN)μ00N),b=(c0NTr0), where r is the N×J matrix of discrete returns, 0N denotes the row vector of length N with zero entries and eN – the row vector of length N with all entries equal 1. We take x^=(wb,0,0N) and B=(diag(eJ)00diag(εN+1)), where εk is the row vector of length k with all entries equal ε.

With the above definitions, the solution obtained by the central path projection algorithm is an optimal solution to problem (32) with constraint (33).

For the mean-LSAD problem with X={u:u0}, the linear program is (34){n=1Npnynmin,yn0,n=1,,N,yn+r^nTu0,n=1,,N,uTμr0,u0. To reduce this problem to the standard form appropriate for the central path projection algorithm, we define x=(u,y), c=(0J,p), x^=(wb,0N) and matrices A=(r^diag(eN)μ0N),b=(0NTr0),B=(diag(eJ)00diag(εN)), where r^ is the N×J matrix of discrete centered returns. Running the central path projection algorithm with the above defined vectors and matrices, we obtain an optimal solution to problem (34) with constraint (33).

5 Numerical examples

The algorithms described in Sections Benders decomposition and Projection algorithm for portfolio optimization are implemented in R in the package PortfolioOptim. The package offers two public functions:

For solving LP subproblems in BDportfolio_optim, we use function Rglpk_solve_LP from the R-package Rglpk. All computations are carried out on a computer with Intel Core i5-5200U processor with 8 GB RAM running Linux operating system.

Benders decomposition algorithm

It has been already observed by that a good discrete approximation of a continuous distribution of returns requires large samples. The size of the sample depends on the goal of optimization procedure. It has been shown in that the value of the objective function is accurately estimated with samples of size 1000020000. Our experience shows that estimation of the optimal portfolio weights requires much larger samples.

Consider first the model used in : they provide the vector of mean μ and the covariance matrix Σ for monthly returns of 5 assets (MSCI.CH, MSCI.E, MSCI.W, Pictet.Bond and JPM.Global), and assume that the asset returns have joint normal distribution. We represent this distribution by a random sample of size N and solve the optimization problem (18) in which we impose additional constrains taking X={u:k=1Kuk=1,u0}. In all computations we take the target portfolio return r0=0.005 and the CVaR confidence level α=0.95.

First we initialize computations with data.

library(mvtnorm)
library(PortfolioOptim)

generate\_data\_normal <- function (means, covmat, num)
\{
    k <- ncol(covmat)
    sim\_data <- rmvnorm  (n=num, mean = means, sigma=covmat)
    sim\_data <- matrix(num, k, data = sim\_data) 
    colnames(sim_data) <- colnames(covmat)
    prob <- matrix(1/num,num,1)
    mod\_returns <- cbind(sim\_data, prob)   
    return (mod\_returns)
\}

prepare\_data\_KM <- function ()
\{
    sample\_cov <- matrix(5,5, data = c( 0.003059 ,  0.002556 , 0.002327 , 0.000095 , 0.000533,
 0.002556 ,  0.003384 , 0.002929 ,  0.000032 ,  0.000762,
 0.002327 ,  0.002929 , 0.003509 ,  0.000036 ,  0.000908,
 0.000095 ,  0.000032 , 0.000036 ,  0.000069  , 0.000048 ,
 0.000533 ,  0.000762 , 0.000908  , 0.000048  , 0.000564))
    sample\_mean <- c( 0.007417,  0.005822,   0.004236,  0.004231,   0.005534)
    colnames(sample\_cov) <- c("MSCI.CH", "MSCI.E", "MSCI.W", "Pictet.Bond", "JPM.Global")
    return(list( sample\_mean = sample\_mean, sample\_cov = sample\_cov))
\}

We perform two tests (both tests are run simultaneously). In the first test, we compare computational time for different sample sizes. In the second test, we compare the accuracy of the obtained optimal portfolios. We perform computations for different sample sizes; for each sample size we generate 10 independent samples and assess the mean and the variance of the running time and portfolio weights. The R code for these tests is as follows.

data\_nA <- prepare\_data\_KM()
sample\_cov <- data\_nA\$sample\_cov 
sample\_mean <- data\_nA\$sample\_mean 
k <- ncol(sample\_cov)
a0 <- rep(1,k)
Aconstr <- rbind(a0,-a0)
bconstr <- c(1+1e-8, -1+1e-8)
lbound <- rep(0,k)
ubound <- rep(1,k)
R0 = 0.005

ET <- NULL
weights <- NULL 
repetition = 10
sample\_size = 10 000  \# also run for 100 000 and 1 000 0000

ptm <- proc.time()[3] 
  for (i in 1:repetition )\{
    mod\_returns <- generate\_data\_normal (sample\_mean, sample\_cov, sample\_size)
    res  <-  BDportfolio\_optim (mod\_returns, R0, risk = "CVAR",  alpha = 0.95, 
            Aconstr, bconstr,lbound, ubound, maxiter = 200, tol = 1e-10 )
    ET <- c(ET, proc.time()[3] - ptm)
    ptm <- proc.time()[3]        
    weights <- rbind(weights, t(res\$theta))
  \}
  
cat(''running time and its standard deviation \(\backslash\)n'')
print(mean(ET))
print(sqrt(var(ET)))    

cat(''optimal portfolio and confidence intervals of its weights \(\backslash\)n'')
print((colMeans(weights))*100)
print(sqrt(apply(weights, 2, var))*100*4/sqrt(repetition))

Since standard LP solvers (used for comparison by ) are too memory demanding to run on our hardware, we report computational time only for our BDportfolio_optim function. Table 1 contains the average running times for 10 different realizations together with the standard deviations.

Table 1: Averaged running time (in sec.) and its standard deviation for computations with different sample sizes.
sample size mean st. dev.
10 000 0.0663 0.0085
100 000 0.4953 0.0363
1 000 000 4.518 0.1086

The average running time for samples of size 10000 is slightly smaller than the value reported by (they give the value 0.088 sec.) but it can be attributed to a slightly faster CPU. It is also visible from Table 1 that even for samples of size 1000000 the computational time is small enough for such sample sizes to be used in practical computations.

The estimates of optimal portfolio weights for different sample sizes are collected in Table 2. The 95% confidence intervals of portfolio weights are reported in brackets. If the difference between portfolio weights is larger that two standard deviations we conclude that the difference is statistically significant (this corresponds to 95% two-sided Student-t test whose critical value for the sample of size 10 is 2.228). The results of Table 2 show that samples of size 10000 are much to small to produce reliable portfolio weights. The values for that sample size are statistically significantly different from the values for 1000000 samples. From the values of confidence intervals of portfolio weights we can conclude that the computation of optimal portfolio weights from samples of size 10000 can give values with an error up to 50% (for all non-zero weights and the sample of size 10000 the confidence intervals are of order of one half of the corresponding weights). This is a clear indication that to obtain reliable values of portfolio weights we have to use samples of size 1000000 or take the average of a large number of repetitions, which is computationally equivalent.

Table 2: Optimal portfolios for different sample sizes. 95% confidence intervals of portfolio weights given in parentheses (all values are in percentage points).
sample size
assets 10000 100000 1000000
MSCI.CH 9.6 11.2 10.9
(4.33) (1.45) (0.39)
MSCI.E 0.0 0.0 0.0
(0.00) (0.00) (0.00)
MSCI.W 0.0 0.0 0.0
(0.00) (0.00) (0.00)
Pictet.Bond 53.6 56.2 56.8
(13.24) (2.33) (0.83)
JPM.Global 36.8 32.6 32.3
(11.41) (2.37) (0.74)

We analyze now the effect of the number of assets in portfolio on the accuracy of computations. To this end, we use the data-set etfdata from the R-package parma. For comparison we take two subsets: etfdata[1:500,1:5](with 5 assets) and etfdata[1:500,1:10] (with 10 assets). We add to our code a new function computing the vector of means and covariance matrix for a data set with k assets and perform computations with k=5 and k=10.

library(parma) 
library(xts)
library(quantmod)

data(etfdata)
prepare\_data <- function (k)
\{
    quotData <- as.xts(etfdata[1:500, 1:k])
    retData = NULL
    for (i in 1:k){ 
      retData <- cbind(retData,weeklyReturn(quotData[,i], type='arithmetic'))
    } 
    colnames(retData) <- colnames(quotData)
    sample_cov <- cov(retData)  
    sample_mean <- colMeans(retData)
    return(list(sample\_mean = sample\_mean, sample\_cov = sample\_cov))    
\}  

  data\_nA <- prepare\_data(5)  \# also run with prepare\_data(10)
  sample\_cov <- data\_nA\$sample\_cov 
  sample\_mean <- data\_nA\$sample\_mean 
  k <- ncol(sample\_cov)
  a0 <- rep(1,k)
  Aconstr <- rbind(a0,-a0)
  bconstr <- c(1+1e-8, -1+1e-8)
  lbound <- rep(0,k)
  ubound <- rep(1,k)
  
  R0 = 0.004
  sample\_size = 10 000  \# also run for 100 000 and 1 000 0000
  repetition = 100
  weights <- NULL   
  for (i in 1:repetition )\{
      returns <- generate\_data\_normal(sample\_mean, sample\_cov, sample\_size)  
      res  <-  BDportfolio_optim (returns, R0, risk = "CVAR",  alpha = 0.95, 
              Aconstr, bconstr, lbound, ubound, maxiter = 200, tol = 1e-10 )
      weights <- rbind(weights, t(res\$theta))
  \}
  print(sum(sqrt(apply(weights, 2, var))*100*4/sqrt(repetition))/k)

For each data set we generate samples of size 10000, 100000 and 1000000. For each sample size and data set we compute optimal portfolios. These computations are repeated 100 times and the empirical variances of portfolio weights are calculated. Using these results we compute widths of 95% confidence intervals of portfolio weights. To facilitate comparison of results between 5 and 10 assets we report in Table 3 the average width of 95% confidence interval per asset, i.e. the sum of widths for each asset divided by the number of assets. Notice that there is no significant difference in accuracy between portfolios of 5 and 10 assets. It can be interpreted as a kind of robustness of Benders decomposition algorithm. The results fit almost perfectly to the theoretical picture of the square root dependence of confidence intervals on sample size. The values in Table 3 for 1000000 samples confirm that with that sample size we can get almost perfect estimate of portfolio weights with the average confidence interval of 0.1–0.2% which is more than sufficient for practitioners.

Table 3: The average width of 95% confidence interval of portfolio weights (values are in percentage points).
sample size
assets number 10000 100000 1000000
5 0.97 0.29 0.09
10 1.18 0.42 0.15

 

Projection algorithm

In testing the projection algorithm described in Section Projection algorithm for portfolio optimization the following values of parameters are used δ=0.5, p=0.8, σ=103, b1=0.9 and b2=0.9. The starting point in all computations is calculated by the procedure described in that Section. The algorithm is tested on simulated data generated from a normal distribution using the vectors of means and covariance matrices estimated from the previously used data sets etfdata[1:500,1:5] and etfdata[1:500,1:10] from the package parma. Samples of 125, 250 and 500 returns are generated assuming they are weekly returns (the vector of means and covariance matrix appropriately scaled) which corresponds to 2.5, 5 and 10 years of weekly data. For each sample we compute the vector of portfolio weights using the function PortfolioOptimProjection with tol = 1e-6 and taking as the benchmark the portfolio 1/J, i.e. the portfolio with all weights equal to 1/J, where J is the number of assets in the portfolio. The computation for each sample size and asset set is repeated 10 times.

The code is as follows

data\_nA <- prepare\_data(5) \# also run with prepare\_data(10)
sample\_cov <- data\_nA\$sample\_cov 
sample\_mean <- data\_nA\$sample\_mean
k <- ncol(sample\_cov)
a0 <- rep(1,k)
Aconstr <- rbind(a0,-a0)
bconstr <- c(1+1e-8, -1+1e-8)
lbound <- rep(0,k)
ubound <- rep(1,k)

w\_m = rep(1/k,k)
R0 = 0.005
returns_size =125 \# also run for 250 and 500
ET <- NULL
ptm <- proc.time()[3] 
for (i in 1:10 )\{
    mod\_returns <- generate\_data\_normal(sample\_mean, sample\_cov, returns_size )
    res  <-  PortfolioOptimProjection (mod\_returns, R0 , risk = "CVAR",  alpha = 0.95, 
            w\_m, Aconstr, bconstr,lbound, ubound, 800, tol = 1e-6 )
    ET <- c(ET, proc.time()[3] - ptm)
    ptm <- proc.time()[3]
\}
cat(''Mean running time and its standard deviation \(\backslash\)n'')
print(c( mean(ET), sqrt(var(ET))))

The average running time per one computation and its standard deviation is reported in Table 4.

Table 4: Average running time and its standard deviation for the computation of optimal portfolios by the central path projection algorithm for different number of assets and different sample sizes (all values in sec. of CPU time).
number of assets sample size mean sd. deviation
5 125 6.8618 5.7377
250 35.1910 30.2499
500 134.9312 95.2407
10 125 3.1451 1.6976
250 30.0261 22.5226
500 125.1935 104.7991

Taking into account that the sample of size n corresponds to the matrix A of approximate dimension n×n the results of Table 4 are compatible with the results reported in Table 6.1 of . The observation which is missed in is the standard deviation of the running time. This standard deviation is of the same order of magnitude as the average running time. We can conclude that the computational time depends not only on the size of matrix A but also on the entries, which is particularly surprising since samples are drawn from the same distribution. Of course, the sample of size 125 can be considered as small even for returns of 5 assets. But the sample of size 500 is already quite large. In addition, the averaged running time is increasing with the sample size, however that increase is very irregular. At the same time, the standard deviation behaves also irregularly (similarly as for the averaged running time). All these observations show that the computation time is very sensitive not only to the dimensionality of the problem but also to a particular realization of the data. However we can still conclude that on average the algorithm is efficient for moderate-size portfolio optimization problems. Indeed, already observed that the convergence rate of their algorithm was slow. The same is valid for our extension of their algorithm, which makes the algorithm not practically applicable for portfolio optimization problems when the size of the discrete distribution of returns is larger than 103. There is no clear indication which part of the algorithm has the main effect on slowing down the convergence.

6 Summary

We presented two optimization algorithms for financial portfolios. These algorithms find optimal portfolios in problems when the nonlinear optimization in mean-risk setting can be reformulated as a linear programming problem. The algorithm implementing Benders cuts allows for large samples. Our experience with very large data samples obtained by simulation from a given distribution shows that the algorithm is robust and estimated optimal portfolios are very stable (the standard deviation of portfolio weights is below 0.5 %). The second algorithm based on the central path projection can be useful for small data samples where solutions are not unique and form a multidimensional simplex. Extracting from this simplex a point with the smallest distance to a given benchmark portfolio can be used to improve the decision process in asset management. However, the second algorithm which implements the central path projection requires further analysis. In particular, we would like to make the algorithm’s performance and running time less dependent on the realization of the data.

Acknowledgments

The author gratefully acknowledges financial support from National Science Centre, Poland, project 2014/13/B/HS4/00176.

CRAN packages used

fPortfolio, PortfolioAnalytics, Rglpk, quadprog, DEoptim, GenSA, psoptim, parma, nloptr, PortfolioOptim

CRAN Task Views implied by cited packages

Finance, Optimization

Note

This article is converted from a Legacy LaTeX article using the texor package. The pdf version is the official version. To report a problem with the html, refer to CONTRIBUTE on the R Journal homepage.

Footnotes

    References

    J. F. Benders. Partitioning procedures for solving mixed-variables programming problems. Numerische Mathematik, 4: 238–252, 1962. URL https://doi.org/10.1007/BF01386316.
    A. Ghalanos. parma: Portfolio allocation and risk management applications. 2016. URL https://cran.r-project.org/web/packages/parma/parma.pdf. R package version 1.5-3.
    W. K. Klein Haneveld and M. H. Van der Vlerk. Integrated chance constraints: Reduced forms and an algorithm. Computational Management Science, 3: 245–269, 2006. URL https://doi.org/10.1007/s10287-005-0007-3.
    R. Koenker and I. Mizera. Convex optimization in R. Journal of Statistical Software, 60(5): 1–23, 2014. URL https://doi.org/10.18637/jss.v060.i05.
    H. Konno. Piecewise linear risk function and portfolio optimization. Journal of the Operations Research Society of Japan, 33: 139–156, 1990. URL https://doi.org/10.15807/jorsj.33.139.
    H. Konno, H. Waki and A. Yuuki. Portfolio optimization under lower partial risk measures. Asia-Pacific Financial Markets, 9: 127–140, 2002. URL https://doi.org/10.1023/A:1022238119491.
    H. Konno and H. Yamazaki. Mean-absolute deviation portfolio optimization model and its application to Tokyo stock market. Management Science, 37: 519–531, 1991. URL https://doi.org/10.1287/mnsc.37.5.519.
    A. Künzi-Bay and J. Mayer. Computational aspects of minimizing conditional value at risk. Computational Management Science, 3: 3–27, 2006. URL https://doi.org/10.1007/s10287-005-0042-0.
    R. Mansini, W. Ogryczak and M. G. Speranza. Twenty years of linear programming based portfolio optimization. European Journal of Operational Research, 234: 518–535, 2014. URL https://doi.org/10.1016/j.ejor.2013.08.035.
    B. G. Peterson and P. Carl. Portfolio analysis, including numerical methods for optimization of portfolios. 2015. URL https://cran.r-project.org/web/packages/PortfolioAnalytics/PortfolioAnalytics.pdf. R package version 1.0.3636.
    B. Pfaff. Financial risk modelling and portfolio optimization with r, 2nd ed. John Wiley & Sons, 2016. ISBN 978-1-119-11966-1.
    R. T. Rockafellar and S. Uryasev. Optimization of conditional value-at-risk. Journal of Risk, 2: 21–42, 2000. URL https://doi.org/10.21314/JOR.2000.038.
    R. T. Rockafellar, S. Uryasev and M. Zabarankin. Master funds in portfolio analysis with general deviation measures. Journal of Banking and Finance, 30: 743–778, 2006. URL https://doi.org/10.1016/j.jbankfin.2005.04.004.
    S. Theussl and H. W. Borchers. CRAN task view: Optimization and mathematical programming. CRAN, 2016. URL http://cran.r-project.org/web/views/Optimization.html.
    D. Würtz, Y. Chalabi, W. Chen and A. E. Pfaff. Portfolio optimization with r/rmetrics. Rmetrics Association & Finance Online, 2009.
    Y.-B. Zhao and D. Li. Locating the least 2-norm solution of linear programs via a path-following method. SIAM Journal on Optimization, 12(4): 893–912, 2002. URL https://doi.org/10.1137/S1052623401386368.

    Reuse

    Text and figures are licensed under Creative Commons Attribution CC BY 4.0. The figures that have been reused from other sources don't fall under this license and can be recognized by a note in their caption: "Figure from ...".

    Citation

    For attribution, please cite this work as

    Palczewski, "LP Algorithms for Portfolio Optimization: The PortfolioOptim Package", The R Journal, 2018

    BibTeX citation

    @article{RJ-2018-028,
      author = {Palczewski, Andrzej},
      title = {LP Algorithms for Portfolio Optimization: The PortfolioOptim Package},
      journal = {The R Journal},
      year = {2018},
      note = {https://rjournal.github.io/},
      volume = {10},
      issue = {1},
      issn = {2073-4859},
      pages = {308-327}
    }