Source code for pypfopt.black_litterman

"""
The ``black_litterman`` module houses the BlackLittermanModel class, which
generates posterior estimates of expected returns given a prior estimate and user-supplied
views. In addition, two utility functions are defined, which calculate:

- market-implied prior estimate of returns
- market-implied risk-aversion parameter
"""
import sys
import warnings
import numpy as np
import pandas as pd
from . import base_optimizer


[docs]def market_implied_prior_returns( market_caps, risk_aversion, cov_matrix, risk_free_rate=0.02 ): r""" Compute the prior estimate of returns implied by the market weights. In other words, given each asset's contribution to the risk of the market portfolio, how much are we expecting to be compensated? .. math:: \Pi = \delta \Sigma w_{mkt} :param market_caps: market capitalisations of all assets :type market_caps: {ticker: cap} dict or pd.Series :param risk_aversion: risk aversion parameter :type risk_aversion: positive float :param cov_matrix: covariance matrix of asset returns :type cov_matrix: pd.DataFrame or np.ndarray :param risk_free_rate: risk-free rate of borrowing/lending, defaults to 0.02. You should use the appropriate time period, corresponding to the covariance matrix. :type risk_free_rate: float, optional :return: prior estimate of returns as implied by the market caps :rtype: pd.Series """ mcaps = pd.Series(market_caps) mkt_weights = mcaps / mcaps.sum() # Pi is excess returns so must add risk_free_rate to get return. return risk_aversion * cov_matrix.dot(mkt_weights) + risk_free_rate
[docs]def market_implied_risk_aversion(market_prices, frequency=252, risk_free_rate=0.02): r""" Calculate the market-implied risk-aversion parameter (i.e market price of risk) based on market prices. For example, if the market has excess returns of 10% a year with 5% variance, the risk-aversion parameter is 2, i.e you have to be compensated 2x the variance. .. math:: \delta = \frac{R - R_f}{\sigma^2} :param market_prices: the (daily) prices of the market portfolio, e.g SPY. :type market_prices: pd.Series with DatetimeIndex. :param frequency: number of time periods in a year, defaults to 252 (the number of trading days in a year) :type frequency: int, optional :param risk_free_rate: risk-free rate of borrowing/lending, defaults to 0.02. The period of the risk-free rate should correspond to the frequency of expected returns. :type risk_free_rate: float, optional :raises TypeError: if market_prices cannot be parsed :return: market-implied risk aversion :rtype: float """ if not isinstance(market_prices, (pd.Series, pd.DataFrame)): raise TypeError("Please format market_prices as a pd.Series") rets = market_prices.pct_change().dropna() r = rets.mean() * frequency var = rets.var() * frequency return (r - risk_free_rate) / var
[docs]class BlackLittermanModel(base_optimizer.BaseOptimizer): """ A BlackLittermanModel object (inheriting from BaseOptimizer) contains requires a specific input format, specifying the prior, the views, the uncertainty in views, and a picking matrix to map views to the asset universe. We can then compute posterior estimates of returns and covariance. Helper methods have been provided to supply defaults where possible. Instance variables: - Inputs: - ``cov_matrix`` - np.ndarray - ``n_assets`` - int - ``tickers`` - str list - ``Q`` - np.ndarray - ``P`` - np.ndarray - ``pi`` - np.ndarray - ``omega`` - np.ndarray - ``tau`` - float - Output: - ``posterior_rets`` - pd.Series - ``posterior_cov`` - pd.DataFrame - ``weights`` - np.ndarray Public methods: - ``default_omega()`` - view uncertainty proportional to asset variance - ``idzorek_method()`` - convert views specified as percentages into BL uncertainties - ``bl_returns()`` - posterior estimate of returns - ``bl_cov()`` - posterior estimate of covariance - ``bl_weights()`` - weights implied by posterior returns - ``portfolio_performance()`` calculates the expected return, volatility and Sharpe ratio for the allocated portfolio. - ``set_weights()`` creates self.weights (np.ndarray) from a weights dict - ``clean_weights()`` rounds the weights and clips near-zeros. - ``save_weights_to_file()`` saves the weights to csv, json, or txt. """
[docs] def __init__( self, cov_matrix, pi=None, absolute_views=None, Q=None, P=None, omega=None, view_confidences=None, tau=0.05, risk_aversion=1, **kwargs ): """ :param cov_matrix: NxN covariance matrix of returns :type cov_matrix: pd.DataFrame or np.ndarray :param pi: Nx1 prior estimate of returns, defaults to None. If pi="market", calculate a market-implied prior (requires market_caps to be passed). If pi="equal", use an equal-weighted prior. :type pi: np.ndarray, pd.Series, optional :param absolute_views: a colleciton of K absolute views on a subset of assets, defaults to None. If this is provided, we do not need P, Q. :type absolute_views: pd.Series or dict, optional :param Q: Kx1 views vector, defaults to None :type Q: np.ndarray or pd.DataFrame, optional :param P: KxN picking matrix, defaults to None :type P: np.ndarray or pd.DataFrame, optional :param omega: KxK view uncertainty matrix (diagonal), defaults to None Can instead pass "idzorek" to use Idzorek's method (requires you to pass view_confidences). If omega="default" or None, we set the uncertainty proportional to the variance. :type omega: np.ndarray or Pd.DataFrame, or string, optional :param view_confidences: Kx1 vector of percentage view confidences (between 0 and 1), required to compute omega via Idzorek's method. :type view_confidences: np.ndarray, pd.Series, list, optional :param tau: the weight-on-views scalar (default is 0.05) :type tau: float, optional :param risk_aversion: risk aversion parameter, defaults to 1 :type risk_aversion: positive float, optional :param market_caps: (kwarg) market caps for the assets, required if pi="market" :type market_caps: np.ndarray, pd.Series, optional :param risk_free_rate: (kwarg) risk_free_rate is needed in some methods :type risk_free_rate: float, defaults to 0.02 """ if sys.version_info[1] == 5: # if python 3.5 warnings.warn( "When using python 3.5 you must explicitly construct the Black-Litterman inputs" ) # Keep raw dataframes self._raw_cov_matrix = cov_matrix #  Initialise base optimiser if isinstance(cov_matrix, np.ndarray): self.cov_matrix = cov_matrix super().__init__(len(cov_matrix), list(range(len(cov_matrix)))) else: self.cov_matrix = cov_matrix.values super().__init__(len(cov_matrix), cov_matrix.columns) #  Sanitise inputs if absolute_views is not None: self.Q, self.P = self._parse_views(absolute_views) else: self._set_Q_P(Q, P) self._set_risk_aversion(risk_aversion) self._set_pi(pi, **kwargs) self._set_tau(tau) # Make sure all dimensions work self._check_attribute_dimensions() self._set_omega(omega, view_confidences) # Private intermediaries self._tau_sigma_P = None self._A = None self.posterior_rets = None self.posterior_cov = None
def _parse_views(self, absolute_views): """ Given a collection (dict or series) of absolute views, construct the appropriate views vector and picking matrix. The views must be a subset of the tickers in the covariance matrix. {"AAPL": 0.20, "GOOG": 0.12, "XOM": -0.30} :param absolute_views: absolute views on asset performances :type absolute_views: dict, pd.Series """ if not isinstance(absolute_views, (dict, pd.Series)): raise TypeError("views should be a dict or pd.Series") # Coerce to series views = pd.Series(absolute_views) k = len(views) Q = np.zeros((k, 1)) P = np.zeros((k, self.n_assets)) for i, view_ticker in enumerate(views.keys()): try: Q[i] = views[view_ticker] P[i, list(self.tickers).index(view_ticker)] = 1 except ValueError: #  Could make this smarter by just skipping raise ValueError("Providing a view on an asset not in the universe") return Q, P def _set_Q_P(self, Q, P): if isinstance(Q, (pd.Series, pd.DataFrame)): self.Q = Q.values.reshape(-1, 1) elif isinstance(Q, np.ndarray): self.Q = Q.reshape(-1, 1) else: raise TypeError("Q must be an array or dataframe") if isinstance(P, pd.DataFrame): self.P = P.values elif isinstance(P, np.ndarray): self.P = P elif len(self.Q) == self.n_assets: # If a view on every asset is provided, P defaults # to the identity matrix. self.P = np.eye(self.n_assets) else: raise TypeError("P must be an array or dataframe") def _set_pi(self, pi, **kwargs): if pi is None: warnings.warn("Running Black-Litterman with no prior.") self.pi = np.zeros((self.n_assets, 1)) elif isinstance(pi, (pd.Series, pd.DataFrame)): self.pi = pi.values.reshape(-1, 1) elif isinstance(pi, np.ndarray): self.pi = pi.reshape(-1, 1) elif pi == "market": if "market_caps" not in kwargs: raise ValueError( "Please pass a series/array of market caps via the market_caps keyword argument" ) market_caps = kwargs.get("market_caps") risk_free_rate = kwargs.get("risk_free_rate", 0) self.pi = market_implied_prior_returns( market_caps, self.risk_aversion, self.cov_matrix, risk_free_rate ).reshape(-1, 1) elif pi == "equal": self.pi = np.ones((self.n_assets, 1)) / self.n_assets else: raise TypeError("pi must be an array or series") def _set_tau(self, tau): if tau <= 0 or tau > 1: raise ValueError("tau should be between 0 and 1") self.tau = tau def _set_risk_aversion(self, risk_aversion): if risk_aversion <= 0: raise ValueError("risk_aversion should be a positive float") self.risk_aversion = risk_aversion def _set_omega(self, omega, view_confidences): if isinstance(omega, pd.DataFrame): self.omega = omega.values elif isinstance(omega, np.ndarray): self.omega = omega elif omega == "idzorek": if view_confidences is None: raise ValueError( "To use Idzorek's method, please supply a vector of percentage " "confidence levels for each view." ) if not isinstance(view_confidences, np.ndarray): try: view_confidences = np.array(view_confidences).reshape(-1, 1) assert view_confidences.shape[0] == self.Q.shape[0] assert np.issubdtype(view_confidences.dtype, np.number) except AssertionError: raise ValueError( "view_confidences should be a numpy 1D array or vector with the same length " "as the number of views." ) self.omega = BlackLittermanModel.idzorek_method( view_confidences, self.cov_matrix, self.pi, self.Q, self.P, self.tau, self.risk_aversion, ) elif omega is None or omega == "default": self.omega = BlackLittermanModel.default_omega( self.cov_matrix, self.P, self.tau ) else: raise TypeError("self.omega must be a square array, dataframe, or string") K = len(self.Q) assert self.omega.shape == (K, K), "omega must have dimensions KxK" def _check_attribute_dimensions(self): """ Helper method to ensure that all of the attributes created by the initialiser have the correct dimensions, to avoid linear algebra errors later on. :raises ValueError: if there are incorrect dimensions. """ N = self.n_assets K = len(self.Q) assert self.pi.shape == (N, 1), "pi must have dimensions Nx1" assert self.P.shape == (K, N), "P must have dimensions KxN" assert self.cov_matrix.shape == (N, N), "cov_matrix must have shape NxN"
[docs] @staticmethod def default_omega(cov_matrix, P, tau): """ If the uncertainty matrix omega is not provided, we calculate using the method of He and Litterman (1999), such that the ratio omega/tau is proportional to the variance of the view portfolio. :return: KxK diagonal uncertainty matrix :rtype: np.ndarray """ return np.diag(np.diag(tau * P @ cov_matrix @ P.T))
[docs] @staticmethod def idzorek_method(view_confidences, cov_matrix, pi, Q, P, tau, risk_aversion=1): """ Use Idzorek's method to create the uncertainty matrix given user-specified percentage confidences. We use the closed-form solution described by Jay Walters in The Black-Litterman Model in Detail (2014). :param view_confidences: Kx1 vector of percentage view confidences (between 0 and 1), required to compute omega via Idzorek's method. :type view_confidences: np.ndarray, pd.Series, list,, optional :return: KxK diagonal uncertainty matrix :rtype: np.ndarray """ view_omegas = [] for view_idx in range(len(Q)): conf = view_confidences[view_idx] if conf < 0 or conf > 1: raise ValueError("View confidences must be between 0 and 1") # Special handler to avoid dividing by zero. # If zero conf, return very big number as uncertainty if conf == 0: view_omegas.append(1e6) continue P_view = P[view_idx].reshape(1, -1) alpha = (1 - conf) / conf # formula (44) omega = tau * alpha * P_view @ cov_matrix @ P_view.T # formula (41) view_omegas.append(omega.item()) return np.diag(view_omegas)
[docs] def bl_returns(self): """ Calculate the posterior estimate of the returns vector, given views on some assets. :return: posterior returns vector :rtype: pd.Series """ if self._tau_sigma_P is None: self._tau_sigma_P = self.tau * self.cov_matrix @ self.P.T # Solve the linear system Ax = b to avoid inversion if self._A is None: self._A = (self.P @ self._tau_sigma_P) + self.omega b = self.Q - self.P @ self.pi post_rets = self.pi + self._tau_sigma_P @ np.linalg.solve(self._A, b) return pd.Series(post_rets.flatten(), index=self.tickers)
[docs] def bl_cov(self): """ Calculate the posterior estimate of the covariance matrix, given views on some assets. Based on He and Litterman (2002). It is assumed that omega is diagonal. If this is not the case, please manually set omega_inv. :return: posterior covariance matrix :rtype: pd.DataFrame """ if self._tau_sigma_P is None: self._tau_sigma_P = self.tau * self.cov_matrix @ self.P.T if self._A is None: self._A = (self.P @ self._tau_sigma_P) + self.omega b = self._tau_sigma_P.T M = self.tau * self.cov_matrix - self._tau_sigma_P @ np.linalg.solve(self._A, b) posterior_cov = self.cov_matrix + M return pd.DataFrame(posterior_cov, index=self.tickers, columns=self.tickers)
[docs] def bl_weights(self, risk_aversion=None): r""" Compute the weights implied by the posterior returns, given the market price of risk. Technically this can be applied to any estimate of the expected returns, and is in fact a special case of efficient frontier optimisation. .. math:: w = (\delta \Sigma)^{-1} E(R) :param risk_aversion: risk aversion parameter, defaults to 1 :type risk_aversion: positive float, optional :return: asset weights implied by returns :rtype: OrderedDict """ if risk_aversion is None: risk_aversion = self.risk_aversion self.posterior_rets = self.bl_returns() A = risk_aversion * self.cov_matrix b = self.posterior_rets raw_weights = np.linalg.solve(A, b) self.weights = raw_weights / raw_weights.sum() return self._make_output_weights()
[docs] def optimize(self, risk_aversion=None): """ Alias for bl_weights for consistency with other methods. """ return self.bl_weights(risk_aversion)
[docs] def portfolio_performance(self, verbose=False, risk_free_rate=0.02): """ After optimising, calculate (and optionally print) the performance of the optimal portfolio. Currently calculates expected return, volatility, and the Sharpe ratio. This method uses the BL posterior returns and covariance matrix. :param verbose: whether performance should be printed, defaults to False :type verbose: bool, optional :param risk_free_rate: risk-free rate of borrowing/lending, defaults to 0.02. The period of the risk-free rate should correspond to the frequency of expected returns. :type risk_free_rate: float, optional :raises ValueError: if weights have not been calcualted yet :return: expected return, volatility, Sharpe ratio. :rtype: (float, float, float) """ if self.posterior_cov is None: self.posterior_cov = self.bl_cov() return base_optimizer.portfolio_performance( self.weights, self.posterior_rets, self.posterior_cov, verbose, risk_free_rate, )