Unsupervised learning is useful when a dataset contains only features and no measurement of the outcome, or when we want to extract information independent from the outcome. Instead of predicting future outcomes, the goal is to learn an informative representation of the data that is useful for solving another task, including the exploration of a data set. Examples include the identification of topics to summarize documents (see Chapter 14, the reduction of the number of features to lower the risk of overfitting and computational cost for supervised learning, or to group similar observations as illustrated by the use of clustering for asset allocation at the end of this chapter.
Dimensionality reduction and clustering are the main tasks for unsupervised learning:
- Dimensionality reduction transforms the existing features into a new, smaller set while minimizing the loss of information. A broad range of algorithms exists that differ by how they measure the loss of information, whether they apply linear or non-linear transformations or the constraints they impose on the new feature set.
- Clustering algorithms identify and group similar observations or features instead of identifying new features. Algorithms differ in how they define the similarity of observations and their assumptions about the resulting groups.
More specifically, this chapter covers:
- how principal and independent component analysis perform linear dimensionality reduction
- how to apply PCA to identify risk factors and eigen portfolios from asset returns
- how to use non-linear manifold learning to summarize high-dimensional data for effective visualization
- how to use T-SNE and UMAP to explore high-dimensional alternative image data
- how k-Means, hierarchical, and density-based clustering algorithms work
- how to apply agglomerative clustering to build robust portfolios according to hierarchical risk parity
This section covers the curse of dimensionality as well as linear and non-linear dimensionality reduction algorithms.
The number of dimensions of a dataset matter because each new dimension can add signal concerning an outcome. However, there is also a downside known as the curse of dimensionality: as the number of independent features grows while the number of observations remains constant, the average distance between data points also grows, and the density of the feature space drops exponentially. The implications for machine learning are dramatic because prediction becomes much harder when observations are more distant, i.e., different from each other.
The notebook curse_of_dimensionality simulates how the average and minimum distances between data points increase as the number of dimensions grows.
Linear dimensionality reduction algorithms compute linear combinations that translate, rotate, and rescale the original features to capture significant variation in the data, subject to constraints on the characteristics of the new features.
This section introduces these two algorithms and then illustrates how to apply PCA to asset returns to learn risk factors from the data, and to build so-called eigen portfolios for systematic trading strategies.
- Dimension Reduction: A Guided Tour, Chris J.C. Burges, Foundations and Trends in Machine Learning, January 2010
PCA finds principal components as linear combinations of the existing features and uses these components to represent the original data. The number of components is a hyperparameter that determines the target dimensionality and needs to be equal to or smaller than the number of observations or columns, whichever is smaller.
The notebook pca_key_ideas visualizes principal components in 2D and 3D.
PCA aims to capture most of the variance in the data, to make it easy to recover the original features, and that each component adds information. It reduces dimensionality by projecting the original data into the principal component space. PCA makes several assumptions that are important to keep in mind. These include:
- high variance implies a high signal-to-noise ratio
- the data is standardized so that the variance is comparable across features
- linear transformations capture the relevant aspects of the data, and
- higher-order statistics beyond the first and second moment do not matter, which implies that the data has a normal distribution
The emphasis on the first and second moments align with standard risk/return metrics, but the normality assumption may conflict with the characteristics of market data.
The notebook the_math_behind_pca illustrate the computation of principal components.
- Mixtures of Probabilistic Principal Component Analysers, Michael E. Tipping and Christopher M. Bishop, Neural Computation 11(2), pp 443–482. MIT Press
- Finding Structure with Randomness: Probabilistic Algorithms for Constructing Approximate Matrix Decompositions, N. Halko†, P. G. Martinsson, J. A. Tropp, SIAM REVIEW, Vol. 53, No. 2, pp. 217–288
- Relationship between SVD and PCA. How to use SVD to perform PCA?, excellent technical CrossValidated StackExchange answer with visualization
PCA is useful for algorithmic trading in several respects, including the data-driven derivation of risk factors by applying PCA to asset returns, and the construction of uncorrelated portfolios based on the principal components of the correlation matrix of asset returns.
In Chapter 07 - Linear Models, we explored risk factor models used in quantitative finance to capture the main drivers of returns. These models explain differences in returns on assets based on their exposure to systematic risk factors and the rewards associated with these factors.
In particular, we explored the Fama-French approach that specifies factors based on prior knowledge about the empirical behavior of average returns, treats these factors as observable, and then estimates risk model coefficients using linear regression. An alternative approach treats risk factors as latent variables and uses factor analytic techniques like PCA to simultaneously estimate the factors and how the drive returns from historical returns.
The notebook pca_and_risk_factor_models demonstrates how this method derives factors in a purely statistical or data-driven way with the advantage of not requiring ex-ante knowledge of the behavior of asset returns.
- Characteristics Are Covariances: A Unified Model of Risk and Return, Kelly, Pruitt and Su, NBER, 2018
- Statistical Arbitrage in the U.S. Equities Market, Marco Avellaneda and Jeong-Hyun Lee, 2008
- Independent Component Analysis: Algorithms and Applications, Aapo Hyvärinen and Erkki Oja, Neural Networks, 2000
- Independent Components Analysis, CS229 Lecture Notes, Andrew Ng
- Common factors in prices, order flows, and liquidity, Hasbrouck and Seppi, Journal of Financial Economics, 2001
- Volatility Modelling of Multivariate Financial Time Series by Using ICA-GARCH Models, Edmond H. C. Wu, Philip L. H. Yu, in: Gallagher M., Hogan J.P., Maire F. (eds) Intelligent Data Engineering and Automated Learning - IDEAL 2005
- The Prediction Performance of Independent Factor Models, Chan, In: proceedings of the 2002 IEEE International Joint Conference on Neural Networks
- An Overview of Independent Component Analysis and Its Applications, Ganesh R. Naik, Dinesh K Kumar, Informatica 2011
The manifold hypothesis emphasizes that high-dimensional data often lies on or near a lower-dimensional non-linear manifold that is embedded in the higher dimensional space.
The notebook manifold_learning_intro contains several exampoles, including the two-dimensional swiss roll that illustrates such a topological structure. Manifold learning aims to find the manifold of intrinsic dimensionality and then represent the data in this subspace. A simplified example uses a road as one-dimensional manifolds in a three-dimensional space and identifies data points using house numbers as local coordinates.
This section uses the following datasets:
Several techniques approximate a lower dimensional manifold. One example is locally-linear embedding (LLE) that was developed in 2000 by Sam Roweis and Lawrence Saul.
The notebook manifold_learning_lle demonstrates how it ‘unrolls’ the swiss roll. For each data point, LLE identifies a given number of nearest neighbors and computes weights that represent each point as a linear combination of its neighbors. It finds a lower-dimensional embedding by linearly projecting each neighborhood on global internal coordinates on the lower-dimensional manifold and can be thought of as a sequence of PCA applications.
- Locally Linear Embedding, Sam T. Roweis and Lawrence K. Saul (LLE author website)
t-SNE is an award-winning algorithm developed in 2010 by Laurens van der Maaten and Geoff Hinton to detect patterns in high-dimensional data. It takes a probabilistic, non-linear approach to locating data on several different, but related low-dimensional manifolds. The algorithm emphasizes keeping similar points together in low dimensions, as opposed to maintaining the distance between points that are apart in high dimensions, which results from algorithms like PCA that minimize squared distances.
- Visualizing Data using t-SNE, van der Maaten, Hinton, Journal of Machine Learning Research, 2008
- Visualizing Time-Dependent Data Using Dynamic t-SNE, Rauber, Falcão, Telea, Eurographics Conference on Visualization (EuroVis) 2016
- t-Distributed Stochastic Neighbor Embedding Wins Merck Viz Challenge, Kaggle Blog 2016
- t-SNE: Google Tech Talk, van der Maaten, 2013
- Parametric t-SNE, fast t-SNE implementation using Keras by Kyle McDonald
UMAP) is a more recent algorithm for visualization and general dimensionality reduction. It assumes the data is uniformly distributed on a locally connected manifold and looks for the closest low-dimensional equivalent using fuzzy topology. It uses a neighbors parameter that impacts the result similarly as perplexity above.
It is faster and hence scales better to large datasets than t-SNE, and sometimes preserves global structure than better than t-SNE. It can also work with different distance functions, including, e.g., cosine similarity that is used to measure the distance between word count vectors.
- UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction, Leland McInnes, John Healy, 2018
The notebooks manifold_learning_tsne_umap and manifold_learning_asset_prices demonstrate the usage of both t-SNE and UMAP with various data sets, including equity returns.
Both clustering and dimensionality reduction summarize the data. Dimensionality reduction compresses the data by representing it using new, fewer features that capture the most relevant information. Clustering algorithms, in contrast, assign existing observations to subgroups that consist of similar data points.
Clustering can serve to better understand the data through the lens of categories learned from continuous variables. It also permits automatically categorizing new objects according to the learned criteria. Examples of related applications include hierarchical taxonomies, medical diagnostics, or customer segmentation. Alternatively, clusters can be used to represent groups as prototypes, using e.g. the midpoint of a cluster as the best representatives of learned grouping. An example application includes image compression.
Clustering algorithms differ with respect to their strategy of identifying groupings:
- Combinatorial algorithms select the most coherent of different groupings of observations
- Probabilistic modeling estimates distributions that most likely generated the clusters
- Hierarchical clustering finds a sequence of nested clusters that optimizes coherence at any given stage
Algorithms also differ by the notion of what constitutes a useful collection of objects that needs to match the data characteristics, domain and the goal of the applications. Types of groupings include:
- Clearly separated groups of various shapes
- Prototype- or center-based, compact clusters
- Density-based clusters of arbitrary shape
- Connectivity- or graph-based clusters
Important additional aspects of a clustering algorithm include whether
- it requires exclusive cluster membership,
- makes hard, i.e., binary, or soft, probabilistic assignment, and
- is complete and assigns all data points to clusters.
The notebook clustering_algos compares the clustering results for several algorithm using toy dataset designed to test clustering algorithms.
k-Means is the most well-known clustering algorithm and was first proposed by Stuart Lloyd at Bell Labs in 1957.
The algorithm finds K centroids and assigns each data point to exactly one cluster with the goal of minimizing the within-cluster variance (called inertia). It typically uses Euclidean distance but other metrics can also be used. k-Means assumes that clusters are spherical and of equal size and ignores the covariance among features.
The notebooks kmeans_implementation demonstrates how the k-Means algorithm works.
Cluster quality metrics help select among alternative clustering results. The notebook kmeans_evaluation illustrates how to evaluate clustering quality using inertia and the silhouette score.
Hierarchical clustering avoids the need to specify a target number of clusters because it assumes that data can successively be merged into increasingly dissimilar clusters. It does not pursue a global objective but decides incrementally how to produce a sequence of nested clusters that range from a single cluster to clusters consisting of the individual data points.
While hierarchical clustering does not have hyperparameters like k-Means, the measure of dissimilarity between clusters (as opposed to individual data points) has an important impact on the clustering result. The options differ as follows:
- Single-link: distance between nearest neighbors of two clusters
- Complete link: maximum distance between respective cluster members
- Group average
- Ward’s method: minimize within-cluster variance
The notebook hierarchical_clusterin demonstrates how this algorithm works, and how to visualize and evaluate the results.
Density-based clustering algorithms assign cluster membership based on proximity to other cluster members. They pursue the goal of identifying dense regions of arbitrary shapes and sizes. They do not require the specification of a certain number of clusters but instead rely on parameters that define the size of a neighborhood and a density threshold.
The notebook density_based_clustering demonstrates how DBSCAN and hierarchical DBSCAN work.
Gaussian mixture models (GMM) are a generative model that assumes the data has been generated by a mix of various multivariate normal distributions. The algorithm aims to estimate the mean & covariance matrices of these distributions.
It generalizes the k-Means algorithm: it adds covariance among features so that clusters can be ellipsoids rather than spheres, while the centroids are represented by the means of each distribution. The GMM algorithm performs soft assignments because each point has a probability to be a member of any cluster.
The notebook gaussian_mixture_models demonstrates the application of a GMM clustering model.
The key idea of hierarchical risk parity (HRP) is to use hierarchical clustering on the covariance matrix to be able to group assets with similar correlations together and reduce the number of degrees of freedom by only considering 'similar' assets as substitutes when constructing the portfolio.
The notebook hrp and the python files in subfolder hierarchical_risk_parity illustrate its application.
- Building Diversified Portfolios that Outperform Out-of-Sample, Lopez de Prado, Journal of Portfolio Management, 2015
- Hierarchical Clustering Based Asset Allocation, Raffinot 2016
- Visualizing the Stock Market Structure