Contents

Contents

In this article

- Editor
- 30 min read

While economic information undeniably wields a significant and widespread influence on financial markets, the systematic incorporation of macroeconomic data into trading strategies has thus far been limited. This reflects skepticism towards economic theory and serious data problems, such as revisions, distortions, calendar effects, and, generally, the lack of point-in-time formats. However, the emergence of industry-wide quantamental indicators and the rise of statistical learning methods in financial markets make macroeconomic information more practical and powerful. Successful demonstrations of statistical learning and macro-quantamental indicators have been achieved, with various machine learning techniques poised to further improve the utilization of economic information.

The principal case for incorporating macroeconomic information into trading strategies has long been compelling. Economic theory shows that market prices balance a broader macroeconomic equilibrium and, hence, depend on economic states and shocks. Meanwhile, the __full information efficiency of the broader market is unlikely due to research costs and attention limitations__ (view post here). Discretionary trading, rooted in macroeconomic fundamentals, has a long history and has been the catalyst for numerous successes in the hedge fund industry. Furthermore, trading based on macroeconomic information is not a zero-sum game. Trading profits are not solely derived from the losses of others but are also paid out of the economic gains from a faster and smoother alignment of market prices with economic conditions. Therefore, technological advancements in this field can increase the value generation or “alpha” of the asset management industry overall (view post here).

And yet, macroeconomic data have hitherto played a very modest role in systematic trading. This reflects two major obstacles.

- First,
__the relations between economic information and market prices are often indirect and potentially convoluted__. Building trading signals requires good judgment based on experience with macroeconomic theory and data. Alas, macroeconomics is not normally the core strength of portfolio managers or trading system engineers. Meanwhile, economists do not always converge on clear common views. - Second, to use macroeconomic data in trading,
__professionals must wrangle many deficiencies and inconveniences of standard series__:**Sparse history**: Many economic data series, particularly in emerging economies, have only a few decades of history. Whilst this would be abundant in other fields, in macroeconomics, this__only captures a limited number of business cycles and financial crisis events__. Often, this necessitates looking at multiple currency areas simultaneously and stitching together different data series, depending on what markets used to watch in the recent and more distant past.**Revisions of time series**: Standard economic databases store economic time series in their latest revised state. However,__initial and intermediate releases of many economic indicators, such as GDP or business surveys, may have looked very different.__This is not only because the data sources have updated information and changed methods but also because adjustment factors for seasonal and calendar effects, as well as for data outliers, are being modified with hindsight. The information recorded*for the*past is typically not the information that was available*in*the past.**Dual timestamps**: Unlike market data,__economic records have different observation periods and release dates__. The former are the periods when the economic event occurred, and the latter are the dates on which the statistics became public. Standard economic databases only associate values with observation periods.**Distortions**: Almost__all economic data are at least temporarily distorted relative to what they promise to measure__. For example, inflation data are often affected by one-off tax changes and administered price hikes. Production and balance sheet data often reflect disruptions, such as strikes or unseasonal weather. Also, there can be sudden breaks in time series due to changes in methodology. Occasionally, statistics offices have even released plainly incorrect data for political reasons.**Calendar effects**:__Many economic data series are strongly influenced by seasonal patterns, working day numbers, and school holiday schedules__. While some series are calendar-adjusted by the source, others are not. Also, calendar adjustment is typically incomplete and not comparable across countries.**Multicollinearity**: The variations of many economic time series are correlated due to common influences, such as business cycles and financial crises. Oftentimes, a multitude of data all seem to tell the same story. It is typically__necessary to distill latent factors that make up common trends in macro data.__This can be done using domain knowledge, statistical methods, or combinations of these two (view post here).

Generally, data wrangling means transforming raw, irregular data into clean, tidy data sets. In many fields of research, this requires mainly reformatting and relabelling. For macroeconomic trading indicators, the wrangling and preparation of data is a lot more comprehensive:

__Adapting macroeconomic indicators for trading purposes requires transforming activity records into market information states__. Common procedures include [1] stitching different series across time to account for changing availability and convention, [2] combining updates and revisions of time series into “vintage matrixes” as the basis of a single “point-in-time” series, and [3] assigning publication time stamps to the periodic updates and revisions of time series.__Economic information typically involves filters and adjustments__. The parameters of these filters must be estimated sequentially without look-ahead bias. Standard procedures are seasonal, working day, and calendar adjustment (view post here), special holiday pattern adjustment, outlier adjustment, and flexible filtering of volatile series. Seasonal adjustment is still largely the domain of official software, albeit there are modules in R and Python that provide access to these.- Markets often view information through the lens of economists.
__To track economic analyses over time, one must account for changing models, variables, and parameters__. A plausible evolution of economic analysis__can be replicated through machine learning methods__. This point is very important. Conventional econometric models are immutable and not backtestable because they are built with hindsight and do not aim to replicate*perceived*economic trends of the past but*actual*trends*.*Machine learning can simulate changing models, hyperparameters, and model coefficients. One practical approach is “__two-stage supervised learning__” (view post here). The first stage is scouting features. The second stage evaluates candidate models and selects the one that is best at any point in time. Another practical__statistical learning example is the simulation of the results of “nowcasters” over time__(view post here). This method estimates past information states through a three-step approach of (1) variable pre-selection, (2) orthogonalized factor formation, and (3) regression-based prediction.

News and comments are major drivers for asset prices, probably more so than conventional price and economic data. Yet, no financial professional can read and analyze the vast flow of verbal information. Therefore, comprehensive news analysis is increasingly becoming the domain of __natural language processing, a technology that supports the quantitative evaluation of humans’ natural language__ (view post here). Natural language processing delivers textual information in a structured form that makes it usable for financial market analysis. A range of useful packages is available for extracting and analyzing financial news and comments.

Overall, __statistical programming nowadays allows the construction of quantamental systems__ (view post here). A quantamental system combines customized, high-quality databases and statistical programming outlines in order to systematically investigate relations between market returns and plausible predictors. The term “quantamental” refers to a joint quantitative and fundamental approach to investing.

Macro quantamental indicators __record the market’s information state with respect to macroeconomic activity, balance sheets, and sentiment__. Quantamental indicators are distinct from regular economic time series insofar as they represent information that was available at the time of reference. Consequently, indicator values are comparable to market price data and are well-suited for backtesting trading ideas and implementing algorithmic strategies.

Quantamental indicators increase the market’s macro information efficiency (and trading profits) for two simple reasons:

__Quantamental indicators broaden the scope of easily backtestable and tradable strategy inputs__. Currently, most systematic strategies focus on market data, such as prices and volumes. Quantamental indicators capture critical aspects of the economic environment, such as growth, inflation, profitability, or financial risks, directly and in a format that is similar to price data. Data in this format can be easily combined across macroeconomic concepts and with price data.- Readily available quantamental indicators reduce information costs through scale effects.
__A quantamental system spreads the investment of low-level data wrangling and codifying fundamental domain know-how across many institutions__. For individual managers, developing trading strategies that use fundamentals becomes much more economical. Access to the system removes expenses for data preparation and reduces development time. It also centralizes curation and common-sense oversight. - Finally,
__quantamental indicators reduce moral hazard__in systematic strategy building. Typically, if the production of indicators takes much time and high costs, there is a strong incentive to salvage failed related strategy propositions through “flexible interpretation” and effective data mining.

The main source of macro quantamental information for institutional investors is the J.P. Morgan Macrosynergy Quantamental System (JPMaQS). It is a service that makes it easy to use quantitative-fundamental (“quantamental”) information for financial market trading. With JPMaQS, users can access a wide range of relevant macro quantamental data that are designed for algorithmic strategies, as well as for backtesting macro trading principles in general.

Quantamental indicators are principally based on a two-dimensional data set.

- The first dimension is the timeline of real-time dates or information release dates. It marks the progression of the market’s information state.
- The second dimension is the timeline of observation dates. It describes the history of an indicator for a specific information state.

For any given real-time date, a quantamental indicator is calculated based on the full information state, typically a time series that may be based on other time series and estimates that would be available at or before the real-time date. This information state-contingent time series is called a **data vintage**.

The two-dimensional structure of the data means that, unlike regular time series, quantamental indicators __convey information on two types of changes: changes in reported values and reported changes in values__. The time series of the quantamental indicator itself shows changes in reports arising from updates in the market’s information state. By contrast, quantamental indicators of changes are reported dynamics based on the latest information state alone.

Statistical learning refers to a set of tools or models that help extract insights from datasets, such as macro-quantamental indicators. Not only does statistical learning support the estimation of relations across variables (parameters), but it also governs the choice of models for such estimates (hyperparameters). Moreover, for macro trading, statistical learning has another major benefit: it allows realistic backtesting. Rather than choosing models and features arbitrarily and potentially with hindsight, statistical learning can simulate a rational rules-based choice of method in the past. __Understanding statistical learning is critical in modern financial markets, even for non-quants__(view post here). This is because statistical learning illustrates and replicates how investors’ experiences in markets shape their future behavior.

Within statistical learning pipelines, simple and familiar econometric models can be deployed to simulate point-in-time economic analysis.

**Linear regression**remains the most popular tool for supervised learning in financial markets. It is appropriate if there is a monotonous relation between today’s indicator value and tomorrow’s expected return that can be linearized. Statistical learning based on regression can optimize both model parameters and hyperparameters sequentially and produce signals based on whichever model has predicted returns best up to a point in time (view post here). In the macro trading space,__mixed data sampling (MIDAS) regressions are a useful method for nowcasting economic trends and financial market variables__, such as volatility (view post here). This type of regression allows combining time series of different frequencies and limits the number of parameters that need to be estimated. Linear regression does not only support estimates for the relationship between signals and returns but also for the relationship between contract returns and macro factors. The latter allows immunizing strategy returns against unwanted macro influences (view post here).**Structural vector autoregression**(SVAR) is a practical model class that captures the evolution of a set of linearly related observable time series variables, such as economic data or asset prices. SVAR assumes that all variables depend in fixed proportion on past values of the set and new structural shocks. The method is useful for macro trading strategies (view post here) because it__helps identify specific interpretable markets and macro shocks__(view post here). For example, SVAR can identify short-term policy, growth, or inflation expectation shocks. Once a shock is identified, it can be used for trading in two ways.- First, one can compare the type of shock implied by markets with the actual news flow and detect fundamental inconsistencies.
- Second, different types of shocks may entail different types of subsequent asset price dynamics and, hence, form a basis for systematic strategies.

- Another useful set of models tackles
**dimension reduction**. This refers to__methods that condense the bulk of the information of many macroeconomic time series into a small set__with the most important information for investors. In macroeconomics, there are many related data series that have only limited incremental relevant information value. Cramming all of them into a prediction model undermines estimation stability and transparency. There are three popular types of statistical dimension reduction methods.- The first type of dimension reduction
__selects a subset of “best” explanatory variables by means of regularization__, i.e., reducing coefficient values by penalizing coefficient magnitudes in the optimization function applied for statistical fit. Penalty functions that are linear in individual coefficient values can set some of them to zero. Classic methods of this type are Lasso and Elastic Net (view post here). - The second type
__selects a small set of latent background factors of__all explanatory variables and then uses these background factors for prediction. This is the basic idea behind static and dynamic**factor models**. Factor models are the key technology behind nowcasting in financial markets, a modern approach to monitoring current economic conditions in real-time (view post here). While nowcasting has mostly been used to predict forthcoming data reports, particularly GDP, the underlying factor models can produce a lot more useful information for the investment process, including latent trends, indications of significant changes in such trends, and estimates of the changing importance of various predictor data series (view post here). - The third type
__generates a small set of functions of the original explanatory variables__that historically would have retained their explanatory power and then deploys these for forecasting. This method is called**Sufficient Dimension Reduction**and is more suitable for non-linear relations. (view post here).

- The first type of dimension reduction

Dimension reduction methods not only help to condense information about predictors of trading strategies but also support portfolio construction. In particular, they are suited for detecting latent factors of a broad set of asset prices (view post here). These factors can be used to improve estimates of the covariance structure of these prices and – by extension – to improve the construction of a well-diversified minimum variance portfolio (view post here).

“When data volume swells beyond a human’s ability to discern the patterns in it…we need a new form of intelligence.”

Mansour Raad

Compared with other research fields, data on the relation between macroeconomic developments and modern financial market returns are scant. This reflects the limited history of modern derivatives markets and the rarity of critical macroeconomic events, such as business cycles, policy changes, or financial crises. Superficially, it seems that many data series and data points are available. However, occurrences of major shocks and trends are limited.

The scarcity of major economic events has two major consequences for the application of statistical learning to macro trading strategies:

__Statistical learning must typically use data panels__, i.e., draw on the experience of multiple and diverse macroeconomies over time. Using such two-dimensional datasets calls for special methods of cross-validation and hyperparameter optimization.__Statistical learning for macro trading signals has to accept a steeper (more unfavorable) bias-variance trade-off than other areas of quantitative research__. This means that as one shifts from restrictive to flexible models, the benefits of reduced bias (misspecification) typically come at a high price of enhanced variance (dependence of models on data sets). This reflects the scarcity and seasonality of critical macro events and regimes.

Statistical learning with reasonable and logical priors for model choice can support trading signal generation through sequential optimization based on panel cross-validation for the purpose of trading signal selection, return prediction, and market regime classification (view post here). This approach can broadly be summarized in six steps:

__Specify suitable data frames of features and targets__at the appropriate frequency, typically weekly or monthly. In the features data frame, the columns are indicator categories, and the rows are double indices of currency areas and time periods. The targets are a double-index series of (lagged) target returns.__Define model and hyperparameter grids__. These mark the eligible set of model options over which the statistical learning process optimizes based on cross-validation. It is at this stage that one must apply theoretical priors and restrictions to prevent models from hugging the data of specific economic episodes.__Choose optimization criteria for the cross-validation of models__. The quality of a signal-generating model depends on its ability to predict future target returns and to generate material economic value when applied to positioning. Statistical metrics of these two properties are related but not identical. The choice depends on the characteristics of the trading signal and the objective of the strategy (view post here). In addition to various financial metrics, common machine learning metrics can be employed for model and hyperparameter selection. These include RMSE, MAPE, MAE, balanced accuracy, AUC-PR and F1 score. A special concept is the discriminant ratio (‘D-ratio’), which measures an algorithm’s success in improving risk-adjusted returns versus a related buy-and-hold portfolio (view post here).__Specify the cross-validation splitter for the panel__. Cross-validation is an assessment of the predictive quality of a model based on multiple splits of the data into training and test sets, where each pair is called a “fold.”. Cross-validation splitters for panel data must ascertain the logical cohesion of the training and test sets based on a double index of cross-sections and time periods, ensuring that all sets are sub-panels over common time spans and respecting missing or blacklisted time periods for individual cross-sections (view customized Python classes here).__Perform sequential cross-validation and optimization of models and derive signals__based on concurrent optimal model versions. This process ensures that backtests based on past signals are not contaminated by hindsight regarding the choice of models, features, and hyperparameters.- Finally,
__evaluate the sequentially optimized signals__in terms of predictive power, accuracy, and naïve PnL generation. For example, view a post here on the application or regression-based trading factors for FX trading.

__Machine learning encompasses statistical learning methods but partly automates the construction of forecast models__ through the study of data patterns, the selection of the best functional form for a given level of complexity, and the selection of the best level of complexity for out-of-sample forecasting. Machine learning can add efficiency to classical asset pricing models, such as factor models and macro trading rules, mainly because it is flexible, adaptable, and generalizes knowledge well (view post here). Machine learning is conventionally divided into three main fields: supervised learning, unsupervised learning, and reinforcement learning.

- In
**supervised learning**, one__distinguishes input and output variables and uses an algorithm to learn which function maps the former to the latter__. This covers most statistical learning applications in financial markets. An example is the assessment of whether the change in interest rate differential between two countries can predict the dynamics of their exchange rate. Supervised learning can be divided into regression, where the output variable is a real number, and classification, where the output variable is a category, such as “policy easing” or “policy tightening” for central bank decisions. **Unsupervised learning**only knows input data. Its goal is to__model the underlying structure or distribution of the data in order to learn previously unknown patterns__. Application of unsupervised machine learning techniques includes clustering (partitioning the data set according to similarity), anomaly detection, association mining, and dimension reduction. More specifically, unsupervised learning methods have been proposed to__classify market regimes__, i.e., persistent clusters of market conditions that affect the success of trading factors and strategies (view post here), for example by using the similarity of return correlation matrices across different asset classes (view post here). An advanced method of unsupervised learning is autoencoders, a type of algorithm with the primary purpose of learning an informative representation of the data, as well as a latent presentation that is useful and meaningful.**Reinforcement learning**is a specialized application of machine learning that__interacts with the environment and seeks to improve the way it performs a task so as to maximize its reward__(view post here). The computer employs trial and error. The model designer defines the reward but gives no clues on how to solve the problem. Reinforcement learning holds potential for trading systems because markets are highly complex and quickly changing dynamic systems. Conventional forecasting models have been notoriously inadequate. A self-adaptive approach that can learn quickly from the outcome of actions may be more suitable. Reinforcement learning can benefit trading strategies directly by supporting trading rules and indirectly by supporting the estimation of trading-related indicators, such as real-time growth (view post here).

**Artificial neural networks** have become increasingly practical for (supervised and unsupervised) macro trading research. Neural networks are adaptive machine learning methods that use interconnected layers of neurons. Any given layer of n neurons refers to n learned features. These are passed through a linear map, followed by a one-to-one nonlinear activation function, to form k neurons in the next layer representing a collection of k transformed features. Learning corresponds to finding an optimal collection of trainable weights. Neural networks learn by finding activation function weights and biases through training data.__Recurrent neural networks are a class of neural networks designed to model sequence data such as time series__. Specialized recurrent neural networks have been developed to retain longer memory, particularly LSTM (Long Short-Term Memory) and GRU (Gated Recurrent Unit). The advantage of neural networks is their flexibility in including complex interactions of features, non-linear effects, and various types of non-price information.

Neural networks for financial market trading can be implemented in Python with TensorFlow or PyTorch. For example, neural networks can principally be used to estimate the state of the market on a daily or higher frequency based on an appropriate feature space, i.e., data series that characterize the market (view post here). Also, they have gained prominence for predicting the realized volatility of asset prices (view post here). Beyond, neural networks can be used to detect lagged correlations between different asset prices (view post here) or market price distortions (view post here).

Backtesting refers to calculations of theoretical profits and losses that would have arisen from applying an algorithmic trading strategy in the past. Its function is to assess the quality of a trading strategy in the future. Statistical programming has made backtesting easy. However, its __computational power and convenience can also be corrosive to the investment process due to its tendency to hug temporary patterns__, while data samples for cross-validation are limited. Moreover, the business of algorithmic trading strategies, unfortunately, provides strong incentives for overfitting models and embellishing backtests (view post here). Similarly, academic researchers in the field of trading factors often feel compelled to resort to data mining in order to produce publishable ‘significant’ empirical findings (view post here).

__Good backtests require sound principles and integrity__ (view post here). Sound principles should include [1] formulating a logical economic theory upfront, [2] choosing sample data upfront, [3] keeping the model simple and intuitive, and [4] limiting tryouts when testing ideas. Realistic performance expectations of trading strategies should be based on a range of plausible versions of a strategy, not an optimized one. Bayesian inference works well for that approach, as it estimates both the performance parameters and their uncertainty. The most important principle of all is integrity: aiming to produce good research rather than good backtests and to communicate statistical findings honestly rather than selling them.

__One of the greatest ills of classical market prediction models is exaggerated performance metrics that arise from choosing the model structure with hindsight__. Even if backtests estimate model parameters sequentially and apply them strictly out of sample, the choice of hyperparameters is often made with full knowledge of the history of markets and economies. For example, the type of estimation, the functional form, and – most importantly – the set of considered features are often chosen with hindsight. This hindsight bias can be reduced by sequential hyperparameter tuning or ensemble methods.

- A
**data-driven process for tuning hyperparameters**can partly endogenize model choice. In its simplest form, it__involves three steps: model training, model validation, and method testing__. This process [1] optimizes the parameters of a range of plausible candidate models (hyperparameters) based on a training data set, [2] chooses the best model according to some numerical criterion (such as accuracy or coefficient of determination) based on a separate validation data set, and [3] evaluates the success of the learning method, i.e. the combination of parameter estimation and model selection, by its ability to predict the targets of a further unrelated test set. - An alternative is
**ensemble learning**. Rather than choosing a single model,__ensemble methods combine the decisions of multiple models to improve prediction performance. This combination is governed by a “meta-model.”__For macro trading, this means that the influence of base models is endogenized and data-dependent, and -hence – the overall learning method can be simulated based on the data alone, reducing the hindsight bias from model choice.

Ensemble learning is particularly useful if one uses flexible models, whose estimates vary greatly with the training set because they mitigate these models’ tendency to memorize noise. There are two types of ensemble learning methods:**Heterogeneous ensemble learning**methods train different types of models on the same data set. First, each model makes its prediction. Then, a meta-model aggregates the predictions of the individual models. Preferably, the different models should have different “skills” or strengths. Examples of this approach include the voting classifier, averaging ensembles, and stacking.**Homogeneous ensemble learning**methods use the same model type but are trained on different data. The methods include bootstrap aggregation (bagging), random forests, and popular boosting methods (Adaboost and gradient boosting). Homogeneous ensemble methods have been shown to produce predictive power for credit spread forecasts (view post here), switches between risk parity strategies (paper here), stock returns(paper here), and equity reward-risk timing (view post here).

Integrating transaction costs into the development process of algorithmic trading strategies can be highly beneficial. One can use a “portfolio machine learning method” to that end (view post here).

“If you torture the data long enough, it will confess to anything.”

Ronald Coase

- Editor
- 30 min read

Jupyter Notebook Macro-quantamental scorecards are condensed visualizations of point-in-time economic information for...

- Editor

Jupyter Notebook Regression-based statistical learning is convenient for combining candidate trading factors...

- Editor

Jupyter Notebook Financial markets pay great attention to reported changes in key...

- Editor

Jupyter Notebook of factor calculation Jupyter Notebook of factor analysis Returns of...

- Editor

Trend detection is one type of macroeconomics-based trading strategy (other types are...

- Editor

Jupyter Notebook Macro beta is the sensitivity of a financial contract’s return...

- Editor

Macrosynergy is a London based macroeconomic research and technology company whose founders have developed and employed macro quantamental investment strategies in liquid, tradable asset classes, across many markets and for a variety of different factors to generate competitive, uncorrelated investment returns for institutional investors for over eighteen years. Our quantitative-fundamental (quantamental) computing system tracks a broad range of real-time macroeconomic trends in developed and emerging countries, transforming them into macro systematic quantamental investment strategies. In June 2020 Macrosynergy and J.P. Morgan started a collaboration to scale the quantamental system and to popularize tradable economics across financial markets.

Manage Cookie Consent

To provide the best experiences on **macrosynergy.com**, we use technologies like cookies to store and/or access device information. Consenting to these technologies will allow us to process data such as browsing behavior or unique IDs on this site. Not consenting or withdrawing consent, may adversely affect certain features and functions.

The technical storage or access is strictly necessary for the legitimate purpose of enabling the use of a specific service explicitly requested by the subscriber or user, or for the sole purpose of carrying out the transmission of a communication over an electronic communications network.

The technical storage or access is necessary for the legitimate purpose of storing preferences that are not requested by the subscriber or user.

The technical storage or access that is used exclusively for statistical purposes.
The technical storage or access that is used exclusively for anonymous statistical purposes. Without a subpoena, voluntary compliance on the part of your Internet Service Provider, or additional records from a third party, information stored or retrieved for this purpose alone cannot usually be used to identify you.

The technical storage or access is required to create user profiles to send advertising, or to track the user on a website or across several websites for similar marketing purposes.

- Research
- Macro Information Efficiency
- Quantitative methods for macro information efficiency
- Fundamental Value Estimates
- Macro Trends
- Implicit Subsidies in Financial Markets
- Endogenous Market Risk
- Price Distortions
- Basics of systemic risk management
- Non-Conventional Monetary Policies
- Financial System Risk
- Government Finances

- Quantamental Academy
- About Us