Agent-Based Model Calibration using Machine Learning Surrogates
Taking agent-based models (ABM) closer to the data is an open challenge. This paper explicitly tackles parameter space exploration and calibration of ABMs combining supervised machine-learning and intelligent sampling to build a surrogate meta-model. The proposed approach provides a fast and accurate approximation of model behaviour, dramatically reducing computation time. In that, our machine-learning surrogate facilitates large scale explorations of the parameter-space, while providing a powerful filter to gain insights into the complex functioning of agent-based models. The algorithm introduced in this paper merges model simulation and output analysis into a surrogate meta-model, which facilitates fast and efficient ABM calibration. We successfully apply our approach to the Brock and Hommes (1998) asset pricing model and to the “Island” endogenous growth model (Fagiolo and Dosi, 2003). Performance is evaluated against a relatively large out-of-sample set of parameter combinations, while employing different user-defined statistical tests for output analysis. The results demonstrate the capacity of machine learning surrogates to facilitate fast and precise exploration of agent-based models’ behaviour over their often rugged parameter spaces.
Keywords: agent based model; calibration; machine learning; surrogate; meta-model.
JEL codes: C15, C52, C63.
This paper proposes a novel approach to model calibration and parameter space exploration in agent-based models (ABM), combining supervised machine learning and intelligent sampling in the design of a novel surrogate meta-model.
Agent-based models deal with the study of socio-ecological systems that can be properly conceptualized through a set of micro and macro relationships. One problem with this framework is that the relevant statistical properties for variables of interest are a priori unknown, even to the modeler. Such properties emerge indeed from the repeated interactions among ecologies of heterogeneous, boundedly-rational and adaptive agents.111In the last two decades a variety of ABM have been applied to study many different issues across a broad spectrum of disciplines beyond economics and including ecology (Grimm and Railsback, 2013), health care (Effken et al., 2012), sociology (Macy and Willer, 2002), geography (Brown et al., 2005), bio-terrorism (Carley et al., 2006), medical research (An and Wilensky, 2009), military tactics (Ilachinski, 1997) and many others. See also Squazzoni (2010) for a discussion on the impact of ABM in social sciences, and Fagiolo and Roventini (2012, 2017) for an assessment of macroeconomic policies in agent-based models. As a result, the dynamic properties of the system cannot be studied analytically, the identification of causal mechanisms is not always possible and interactions give rise to the emergence of relationships that cannot simply be deduced by aggregating those of micro variables (Anderson et al., 1972, Tesfatsion and Judd, 2006, Grazzini, 2012, Gallegati and Kirman, 2012). This raises the issue of finding appropriate tools to investigate the emergent behavior of the model with respect to different parameter settings, random seeds, and initial conditions (see also Lee et al., 2015). Once this search is successful, one can safely move to calibration, validation and, finally, employ the model for policy exercises (more on that in Fagiolo and Roventini, 2017). Unfortunately, this procedure is hardly implementable in practice, notably due to large computation times.
Indeed, many ABMs simulate the evolution of a complex system using many parameters and a relatively large number of time steps. In a calibration setting, this rich expressiveness results in a “curse of dimensionality” that lends to an exponential number of critical points along the parameter space, with multiple local maxima, minima and saddle points, which negatively impact the performance of gradient-based search procedures. Exploring model behaviour through all possible parameter combinations (a full factorial exploration) is practically impossible even for small models. Budgetary constraints also restrict our use from multi-objective optimization procedures, such as multimodel optimization or niching (for a review, see e.g. Li et al., 2013; Wong, 2015), and kriging-based procedures due to the large number of evaluations required for these procedures to converge to meaningful interpretations of the model parameter space. 222For example, consider a model with parameters and assume that a single evaluation of the ABM requires seconds on a single compute core (CPU). If one discretizes the parameter space by splitting each dimension into intervals, evaluations would require approximately CPU days to explore. With a finer partition of of say intervals, evaluations would roughly require months, and intervals would require months. Adding a sixth parameter would require more than 10 years. However, if a model is to be useful for policy makers, it must provide timely and accurate insights into the problem. As a result, for computationally expensive models such as ABMs to provide practical insights with their rich expressiveness, they must be efficiently calibrated on a limited budget of evaluations.
Traditionally, three computationally expensive steps are involved in ABM calibration; running the model, measuring calibration quality and locating parameters of interest. As remarked in Grazzini et al. (2017), such steps account for more than half of the time required to estimate ABMs, even for extremely simple models. Recently, kriging (also known as Gaussian processes) has been employed to build surrogate meta-models of ABMs (Salle and Yildizoglu, 2014; Dosi et al., 2016, 2017c, 2017b; Bargigli et al., 2016) to facilitate parameter space exploration and sensitivity analyses. However, kriging cannot be reasonably applied to large scale models with more than 20 parameters even in the linear time extensions proposed in Wilson et al. (2015) and Herlands et al. (2015). Moreover, the smooth surfaces produced by kriging meta-models do not provide an accurate approximation of the rugged parameter spaces characteristic of most ABMs.
In this paper, we explicitly tackle the problem of efficiently exploring the complex parameter space of agent-based models by employing an efficient, adaptive, gradient-free search over the parameter space. The proposed approach exploits both labeled and unlabeled parameter combinations in a semi-supervised manner to build a fast, efficient, machine-learning surrogate mapping a statistic, based on a user-defined measure of fit, and a specific parameterization of the ABM. This procedure results in a dramatic reduction in computation time, while providing an accurate surrogate of the original ABM. This surrogate can then be employed for detailed exploration of the possibly wild parameter space. Moreover, we move towards calibration by identifying parameter combinations that allow the ABM to match user-desired properties.333The interested reader might want to look at van der Hoog (2016) for a broad discussion on possible applications of machine learning algorithms to agent based modelling.
Surrogate meta-models are traditionally employed to approximate or emulate computationally costly experiments or simulation models of complex physical phenomena (see Booker et al., 1999). In particular, surrogates provide a proxy that can be exploited for fast parameter-space exploration and model calibration. Given their speed advantage, surrogates are regularly exploited to locate promising calibration values and gain rapid intuition over a model. Note that the objective is not to return a single optimal parameter, but all parametrizations that positively identify the ABM with user-desired behaviour. Accordingly, if the surrogate approximation error is small, it can be interpreted as an efficient and reasonably good replacement for the original ABM during parameter space exploration and calibration.
Our approach to learning a surrogate occurs over multiple rounds. First, a large “pool” of unlabelled parametrizations are drawn using a standard sampling routine, such as quasi-random Sobol sampling. Next, a very small subset of the pool is randomly drawn without replacement for evaluation in the ABM, making sure to have at least one example of the user-desired behaviour. These points are “labelled” according to the statistic measured on the output generated by the ABM and act as a “seed” set of samples to initialize the surrogate model learned in the first round. This first surrogate is then exploited to predict the label for unlabelled points remaining in the pool. Another very small subset of points are drawn from the pool for evaluation in the agent-based model. Then, over multiple rounds, this process is repeated until a specified budget of evaluations is achieved. In each round, the surrogate directs which unlabelled points are drawn from the pool to maximize the performance of the surrogate learned in the next round. This semi-supervised “active” learning procedure incrementally improves the surrogate model, while maximizing the information gained over the ABM parameter space.444In the Machine Learning jargon supervised learning refers to the task of inferring a function from labeled training data, that is, data that are assigned either a numerical value or a symbol. Semi-supervised learning indicates a setting when there is a small amount of labelled data relatively to unlabelled ones. The term active refers instead refers an algorithm that actively selects which data point to evaluate and, therefore, to label.
The performance of such a procedure crucially depends on the particular surrogate model used in each of the rounds. Here, we automatically tune extremely boosted gradient trees (XGBoost, see Chen and Guestrin, 2016) as our machine-learning surrogate, through automated hyperparameter optimization (Claesen et al., 2014, see), to robustly manage non-linear parameter surfaces and so-called “knife-edge” properties characteristic of ABMs. One particular advantage of this surrogate learning algorithm over kriging is that it does not require the selection of a kernel or to set a prior in advance of the previously mentioned sampling procedure. It also avoids the problem of choosing a summary statistic and acceptance thresholds that comes with likelihood-free approximate Bayesian methods (Grazzini et al., 2017).
As illustrative examples, we apply our procedure to two well known ABMs: the asset pricing model proposed in Brock and Hommes (1998) and the endogenous growth model developed in Fagiolo and Dosi (2003). Despite their relative simplicity, the two models might exhibit multiple equilibria, allow different behavioural attitudes and account for a wide range of dynamics, which crucially depends on their parameters. We find that our machine-learning surrogate is able to efficiently filter out combinations of parameters conveying the output of interest, assess the relative importance of models’ parameters and provide an accurate approximation of the underlying ABM in a negligible amount of time. The advantages in terms of computation cost, hands-free parameter selection and ability to deal with non-linear characteristics of the ABM parameter space of our approach paves the way towards an efficient and user-friendly procedure to parameter space exploration and calibration of agent-based models.
The remaining portions of this paper are organized as follows. Section 2 reviews literature on ABM calibration validation, making the case for surrogate modelling. Section 3 presents our surrogate modelling methodology. Sections 4 and 5 report the results of its application to the asset pricing model proposed in Brock and Hommes (1998) and the growth model developed in Fagiolo and Dosi (2003) respectively. Finally, Section 6 concludes.
2 Calibration and validation of agent-based models: the case for surrogate modelling
As stated in Fagiolo et al. (2007) and Fagiolo and Roventini (2012, 2017), the extreme flexibility of ABMs concerning e.g. various forms of individual behaviour, interaction patterns and institutional arrangements has allowed researchers to explore the positive and normative consequences of departing from the often over-simplifying assumptions characterizing most mainstream analytical models. Recent years have witnessed a trend in macro and financial modeling towards more detailed and richer models, targeting a higher number of stylized facts, and claiming a strong empirical content.555See e.g. Dosi et al. (2010, 2013, 2015); Caiani et al. (2016); Assenza et al. (2015) and Dawid et al. (2014a) on business cycle dynamics, Lamperti et al. (2017) on growth, green transitions and climate change, Dawid et al. (2014b) on regional convergence and Leal et al. (2014) on financial markets. The surveys in Fagiolo and Roventini (2012, 2017) provides a more exhaustive list.
A common theme informing both theoretical analysis and methodological research concerns the relationships between agent-based models and real-world data. Recently, many studies have addressed the problem of estimating and calibrating ABMs. As stated by Chen et al. (2012), ABMs need to move from stage I, i.e. the capability to grow stylized facts in a qualitative sense, to stage II, where appropriate parameter values are selected according to sound econometric techniques. In those cases where the model is sufficiently simple and well behaved, one can derive a closed form solution and estimate the distribution parameters for a specific output of the model (see e.g. Alfarano et al., 2005, 2006; Boswijk et al., 2007). However, when complexity prevents a closed form solution, more sophisticated techniques are required. Amilon (2008) estimates a model of financial markets with 15 parameters (but only 2 or 3 agents) by the method of moments. They report that the model has a high sensitivity to the assumptions made on the noise term and stochastic components. Gilli and Winker (2003) and Winker et al. (2007) introduce an algorithm and a set of statistics leading to the construction of an objective function, which is used to estimate exchange-rate models by indirect inference, pushing them closer to the properties of real data. Franke (2009) refines on this framework and uses the method of simulated moments to estimate 6 parameters of an asset pricing model, while Franke and Westerhoff (2012) propose a model contest for structural stochastic volatility models characterized by few parameters.666See also Grazzini and Richiardi (2015) and Fabretti (2012) for other applications of the same approach Finally, Recchioni et al. (2015) use a simple gradient-based calibration procedure and then test the performance of the model they obtained through out of sample forecasting.
A parallel stream of research has recently focusing on the development of tools to investigate the extent ABM outputs are able to approximate reality (see Marks, 2013; Lamperti, 2017, 2016; Barde, 2016b, a; Guerini and Moneta, 2016). Some of these contributions also offer new measures that can be used to build objective functions in the place of longitudinal moments within an estimation setting (e.g. the GSL-div introduced in Lamperti, 2017). However, a common limitation of both these calibration/estimation and validation exercises lies in their computational time, which is usually extremely high. As discussed in detail by Grazzini et al. (2017), simulating the model is the most computationally expensive step for all these procedures. For instance, in order to train his algorithm, Barde (2016b) needs Monte Carlo (MC) runs each having length of about periods, and many macroeconomic ABMs might take weeks just to perform a single MC exercise of this kind. This explains why the vast majority of previous contributions employ extremely simple ABMs (few parameters, few agents, no stochastic draws) to illustrate their approach, and large macro ABMs are usually poorly validated and calibrated. Hence, using standard statistical techniques, the number of parameters must be minimized to achieve feasible estimation.
From a theoretical perspective, the curse of dimensionality implies that the convergence of any estimator to the true value of a smooth function defined on a high dimensional parameter space is very slow (Weeks, 1995; De Marchi, 2005). Several methods have been introduced in the design of experiments literature to circumvent this problem, but the assumptions of smoothness, linearity and normality do not generally hold for ABMs (see the extensive discussion in Lee et al., 2015).
Unfortunately, recent developments in agent-based macro-economics have led to the development of more and more complex models, which require large sets of parameters to adequately capture the complexity of micro-founded, multi-sector and possibly multi-country phenomena (see Fagiolo and Roventini, 2017, for a recent survey).
In such a setting, neither direct estimations nor global sensitivity analysis (often advocated as a natural approach to ABM exploration, cf. Moss, 2008; Thiele et al., 2014; ten Broeke et al., 2016) seem computationally feasible.
New alternative methods must deal with two issues: reduction in computation time and the design of appropriate criteria for calibration and validation procedures. Our approach shows that such issues can be related in a meaningful way by developing a computational procedure that efficiently trains a surrogate model in order to optimize specific calibration criteria or reproducing statistical relationships between model-generated variables. Our procedure has some similarities to the one of Dawid et al. (2014b), where penalized splines methods are employed to shortcut parameter exploration and unravel the dynamic effects of policies on the economic variables of interest. However, our method especially focuses on computational efficiency and therefore builds on two pillars: surrogate modelling and intelligent sampling.
With respect to surrogate modelling, we extend recent contributions in the economic literature that use kriging to build a surrogate meta-model for ABMs (Salle and Yildizoglu, 2014; Dosi et al., 2017c; Bargigli et al., 2016). One of the primary challenges with kriging-based meta-models is that they cannot efficiently model more than a dozen parameters. This constraint forces modellers to arbitrarily fix a subset of parameters whenever the parameter space is large. Moreover, kriging relies on Gaussian processes (Rasmussen and Williams, 2006; Conti and O’Hagan, 2010), which face serious difficulties when the underlying smoothness assumptions are violated. Modelling the rugged parameter space of ABMs is particularly challenging. In order to overcome these constraints, our meta-modelling approach leverages non-parametric boosted trees from the machine learning literature that do not depend on smoothness assumptions (see Freund et al., 1996; Breiman et al., 1984).
Even the most advanced surrogate modelling algorithm only performs as well as the quality of labelled samples. With respect to ABMs, a labelled sample is a parameter combination and the output of the ABM given this parametrization. Batch sampling, the process of sampling a budget of samples all at once, such as in random sampling, quasi-random sampling (e.g. Sobol sampling), extensions that extend the Sobol sequence to reduce error rates (see Saltelli et al., 2010) and more sophisticated procedures such as Latin-Hypercube sampling are all limited by their one-off nature to sampling. Further, ABM parameters of interest are often rare and represent a small percent of possible parametrizations. Given this imbalanced nature of the sample and the non-negligible computation cost of evaluating ABM parameters, it makes sense to carefully select which parametrizations to evaluate, while exploiting the cheap (almost free) cost of generating unevaluated parametrizations. The problem of sequentially selecting the most informative subset of samples over multiple sampling rounds underlies active learning (see Settles, 2010, for a survey). In particular, given a large pool of unlabelled parametrizations and a fixed evaluation budget, active learning chooses parametrizations from the pool that maximize the generalization or learning performance of the surrogate meta-model.
3 Surrogate modelling methodology
One can represent an agent-based model as a mapping from a set of input parameters into an output set . The set of parameters can be conceived as a multidimensional space spanned by the support of each parameter. The number of parameters in large macroeconomic ABMs generally range up to several dozen. The output set is generally larger, as it corresponds to time-series realizations of a very large number of micro and macro level variables. This rich set of outputs allows a qualitative validation of agent-based models based on their ability to reproduce the statistical properties of empirical data (e.g. non-stationarity of GDP, cross-correlations and relative volatilities of macroeconomic time series), as well as microeconomic distributional characteristics (e.g. distribution of firms’ size, of households’ income, of assets’ returns). Beyond stylized facts, the quantitative validation of an agent-based model also requires the calibration/estimation of the model on a (generally small) set of aggregate variables (e.g. GDP growth rates, inflation and unemployment levels, asset returns etc.).
Without loss of generality, we can represent this quantitative calibration as the determination of input values such that the output satisfies certain calibration conditions, coming from, e.g, a statistical test or the evaluation of a likelihood or loss functions. This is in line, for example, with the method of simulated moments (Gilli and Winker, 2003; Franke and Westerhoff, 2012). We consider two settings:
Binary outcome. In this setting the calibration criterion can be considered as a function, , that maps the ABM output to a binary variable that takes if a certain property of the output (or set of properties) is found, and otherwise. For example, a property that one might want a financial ABM to match is the presence of excess kurtosis in the distribution of returns. This setting leads to what is referred in the machine learning literature as a classification problem.
Real-valued outcome. In this setting the calibration criterion can be considered as a function, , that maps the ABM output to a real valued number providing a quantitative assessment of a certain property of the model. For example, one might want to compute excess kurtosis of simulated data and then compare it to the one obtained from real data. This setting leads to what is referred in the machine learning literature as a regression problem.
To keep consistency with the machine learning terminology, we say that function assigns a label to the parameter vector . Obviously, one would like to find the set of input parameters such that their labels indicate that a chosen condition is met. More formally, we say that is the set of labels indicating that the condition is satisfied. For example, in the case of binary outcome we can say that , which indicates that the chosen property is observed; in the case of real-valued outcome, assuming that expresses the distance between some statistic of the simulated and real data, one might consider or . The latter case reflects exactly the common calibration problem of minimizing some loss function over the parameter space with random restart to avoid ending up in local minima.
We say that a positive calibration is a parameter vector whose label in contained in the set , i.e. . By contrast, a negative calibration is a parameter vector whose label is not contained in .
The problem now is to find all positive calibrations. However, an intensive exploration of the input set is computationally infeasible. As emphasized above, it is crucial to drastically reduce the computation time required to identify positive calibrations.
This paper proposes to train a surrogate model that efficiently approximates the value of using a limited number of input parameters (budget) to evaluate the true ABM. Once the surrogate is trained, it provides an efficient mean of exploring the behaviour of the ABM over the entire parameter space.777Notwithstanding its precision, the surrogate remains an approximation of the original model. We suggest the user, in any case, to identify positive calibrations and further study model’s behaviour therein and in their close neighbourhoods employing the original ABM.
The surrogate training procedure requires three decisions:
Choosing a machine learning algorithm to act as a surrogate for the original ABM, taking care that the assumptions made by the machine learning model do not force unrealistic assumptions on the parameter space;
Selecting a sampling procedure to draw samples from the parameters space in order to train the surrogate;
Selecting a score or criterion that can be used to evaluate the performance of the surrogate.
We prefer to avoid smoothness assumptions and the challenges of selecting a good prior and kernel when using a kriging-based approach (see Rasmussen and Williams, 2006; Ryabko, 2016), so we propose to use extreme gradient boosted trees (XGBoost) (Chen and Guestrin, 2016, see)that form a random ensemble (see Breiman, 2001) of “boosted” (see Freund, 1990; Freund et al., 1996) classification and regression trees (CART) (Breiman et al., 1984, see). This choice endows our surrogate with the ability to learn non-linear “knife-edge” properties, which typically characterize ABM parameter spaces. Sampling should carefully select which parametrizations of an ABM should be evaluated according to the performance of the surrogate. Here, we leverage pool-based active learning according to a pre-specified budget of evaluations888For a review of active learning, see e.g. Settles (2010). The structure of the surrogate, active learning approach and performance criterion are detailed below.
3.1 Structure of the surrogate
Here, we employ an iterative training procedure (see Figure 1) to construct a different surrogate at each of several rounds until we approach a predefined budget of evaluations on the true ABM. At each round an additional parameter vectors is used in the iterative procedure. The budget is set in advance by the user according to a pre-determined, acceptable, computation cost of learning the surrogate. In each round, a surrogate is trained using all available parameter vectors, and their respective labels, which have been aggregated up to that round. Once the budget of evaluations is reached, the final surrogate is ready to be used for parameter space exploration.
Here, we rely on XGBoost (Chen and Guestrin, 2016) as our surrogate learning algorithm. This algorithm sequentially learns an ensemble of classification and regression trees (CART, see Breiman et al., 1984). Figure 2 provides an example of CART tree. Given that the CART trees are represented as functions, the gradient resulting from the ensemble of CART trees can be minimized. Weights are assigned to each of the parameter vectors and “boosted” in the direction of the gradient that minimizes the total loss. Boosting magnifies the importance of difficult-to-learn samples. In each of the subsequent rounds, a new tree is learned over the boosted parameter vectors, incurring an increased penalty according to the boosted weights. Accordingly, trees are learned according to the weight from the previous round. The XGBoost algorithm builds CART trees that are increasingly specialized to handle the particular subset of samples that were difficult to learn up until the current round. A common way to characterize this learning procedure is to consider it as an ensemble of “weak” approximations, that together construct a strong approximation (see Freund, 1990; Freund et al., 1996; Chen and Guestrin, 2016, for more details see).
3.2 Surrogate performance evaluation
A trained surrogate can be used to efficiently explore the behaviour of the ABM over the entire parameter space. Relevant parameter combinations can then be selected for evaluation using the original ABM. Given the desire to avoid evaluating the computationally expensive true ABM, while also identifying positive calibrations, it is critical to maximize the performance of the surrogate to predict these calibrations. Recall that positive calibrations are points in the parameter space that fulfil the specific conditions, specified by an ABM modeller/user. Such conditions might include any test that compares simulated output with real data (e.g. distance between real and simulated moments, a non-parametric test on distribution equality, mean squared prediction errors, etc.) and/or any specific feature the model might generate (e.g. fat tails in a specific distribution, growth rates of any variable above or below a given threshold, correlation patterns among a set of variables, etc.). In the two exercises presented in this paper (cf. Sections 4 and 5 below), both types of conditions are evaluated.
An effective surrogate should maximize the “True Positive Rate” (TPR). Given a set of parameter combinations, the TPR measures the number of positive calibrations predicted by a learned surrogate model against the actual number of positive calibrations possible in the parameter space. Automated hyper-parameter optimization procedures maximize the performance of the machine learning surrogate according to a learning score or metric.999Several procedures exist for tuning machine learning hyper-parameters, see e.g. Feurer et al. (2015). Though our aim is to maximize the TPR of our surrogate, the scores used to train the surrogate depend on the particular form of the output condition. According to the two settings introduced above we distinguish between:
Binary outcome. In this case the output of the calibration condition is discrete, such as Accept/Reject, and a measure of classification ability is needed. Specifically, we aim at maximizing the -score.101010Note that there is “no free lunch” with regard to performance measures, so their choice depends on the problem setting (see e.g. Wolpert, 2002) For a detailed description of the -Score, see e.g. Van Rijsbergen (1979). The -score is an harmonic mean between , which indicates the ratio between true positives and total positives and , which represents the ratio of true positives to predicted ones:
The -score takes a value between and . In terms of Type I and Type II errors, it equates to:
Real-valued outcome. In this case, our aim is to minimize the mean-squared error (MSE),
where the surrogate predicts over evaluation points with a true labelling . We notice that this approach is in line, for instance, with Recchioni et al. (2015).
3.3 Parameter importance
The XGBoost algorithm employed in our surrogate modelling procedure allow us also to perform parameter sensitivity analysis at no costs. In particular, the machine learning algorithm provides an intuitive procedure of assessing the explained variance of the surrogate according to the relative number of times a parameter was “split-on” in the ensemble (for details see e.g. Archer and Kimes, 2008; Louppe et al., 2013; Breiman, 2001). As each tree is constructed according to an optimized splitting of the possible values for a specific parameter vector, and it is increasingly focusing on difficult-to-predict samples, splits dictate the relative importance of parameters in discriminating the output conditions of the ABM. Accordingly, the relative number of splits over a specific parameter provides a quantitative assessment of the surrogate model’s sensitivity to the user-specified conditions specified by the parameter. This also allows a ranking over parameters on the basis of their relative importance in producing model behaviour that satisfies whatever conditions specified by the user. As this procedure is non-parametric, the resulting values should be interpreted as a rank-based statistic. In particular, the relative importance values associated to the number of splits only characterize the specific instantiation of the ensemble. The resulting counts provide insight into the relative performance for each parameter. A changing number of trees would result in a different number of splits for each parameter. As the number of trees approach infinity, the number of splits will converge to the true ratio of splits per parameter by the law of large numbers.
3.4 Training procedure
The primary constraint we face is the limited number of parameter combinations that can be used for model evaluation (budget) without incurring in excessive computational costs. To address this issue, we propose a budgeted online active semi-supervised learning approach that iteratively builds a training set of parameter vectors on which the agent-based model is actually evaluated in order to provide labelled data points for the training of the surrogate. The aim of actively sampling the parameter space is to reduce the discrepancy between the regions that contain a manifold of interest and the function approximation produced by the surrogate model. This semi-supervised learning approach (see e.g. Zhu, 2005; Goldberg et al., 2011) minimizes the number of required evaluations, while improving the performance of the surrogate. Given that evaluated parametrizations are aggregated over several rounds and the stationary nature of the parameter space labels, we can use the convergence results proved in Ross et al. (2011) to provide a guideline on the number of parametrizations to evaluate in each round. In particular, we evaluate parameters per round, with , also ensuring at least one positive calibration in the initial seed round. Noting that the constant can be increased or decreased according to the particular ABM.
Generally, positive calibrations represent a very small percentage of points in the parameter space. For example, the concentration of positive calibrations in both of the ABMs presented in this paper represents less than of the parameters. Our approach exploits this imbalance by iteratively selecting a random subset of positive predicted calibrations over a finite number of rounds. As we use positive predicted calibrations, we exploit semi-supervised learning with the surrogate to select which parameters should be evaluated in the next round. In order to maximize computing speed, the algorithm is initialized with a fixed subset of evaluated parameter combinations that are drawn according to a quasi-random Sobol sampling over the parameter space (Morokoff and Caflisch, 1994).111111Note that in high dimensional spaces, standard design of experiments are computationally costly and show little or no advantage over random sampling (Bergstra and Bengio, 2012; Lee et al., 2015). Further, the number of samples are drawn according to the “total variation” analysis presented in Saltelli et al. (2010). These initial “training” points are then evaluated through the ABM, their labels recorded and finally used to initialize the first surrogate model. Once the surrogate is trained, new parameter combinations are sampled over the entire parameter space and labelled using the surrogate. A random subset of points are then selected from the predicted positive calibrations of the surrogate and evaluated for their true labels using the ABM. Given the convergence rates presented in Ross et al. (2011).
These new points are then added to the training set to train a new surrogate in the next round. This “self-training” procedure exploits the imbalance in the data to incrementally increase true positives, while reducing false positives. Note that this simple self-training procedure may result in no new predicted positives. In this case, the algorithm selects new points according to their predicted binary label entropy, where the latter is defined as the entropy between the predicted positive and negative calibration label probabilities. This incremental procedure continues until the targeted training budget is achieved. The algorithm pseudo-code is presented in Figure 3.
4 Surrogate modelling examples: The Brock and Hommes model
In their seminal contribution, Brock and Hommes (1998) develop an asset pricing model (referred here as B&H), where an heterogeneous population of agents trade generic assets according to different strategies (fundamentalist, chartists, etc.). We briefly introduce the model in (cf. Section 4.1). Then we report the empirical setting (see Section 4.2) and results of our machine learning calibration and exploration exercise (cf. Section 4.3). Recall that the seed of the pseudo-random number generator is fixed and kept constant across runs of the model over different parameter vectors.
4.1 The B&H asset pricing model
There is a population of traders that can invest either in a risk free asset, which is perfectly elastically supplied at a gross return , or in a risky one, which pays an uncertain dividend and has a price denoted by . Wealth dynamics is given by
where and are random variables and is the number of the shares of the risky asset bought at time .
Traders are heterogeneous in terms of their expectations about future prices and dividends and are assumed to be myopic mean-variance maximizers. However, as information about past prices and dividends is publicly available in the market, agents can apply conditional expected value , and variance . The demand for share of agents with expectations of type is computed, solving:
which in turns implies
where controls for agents’ risk aversion and indicates the conditional volatility, assumed to be equal across traders and constant over time. In case of zero supply of outside shares and different trader types, the market equilibrium equation can be written as:
where denotes the share that traders of type hold at time . In presence of homogeneous traders, perfect information and rational expectations, one can derive the no-arbitrage market equilibrium condition:
where the expectation is conditional on all histories of prices and dividends up to time and where indicates the fundamental price. Dividends are independent and identically distributed over time with constant mean, equation (8) has a unique solution where the fundamental price is constant and equal to . In what follows, we will express prices as deviations from the fundamental price, i.e. .
At the beginning of each trading period , agents form expectations about future prices and dividends. Agents are heterogeneous in their forecasts. More specifically, investors believe that, in a heterogeneous world, prices may deviate from the fundamental value by some function depending upon past deviations from the fundamental price. Accordingly, the beliefs about and of agents of type evolve according to:
Many forecasting strategies specifying different trading behaviours and attitudes have been studied in the economic literature, (see e.g. Banerjee, 1992; Brock and Hommes, 1997; Lux and Marchesi, 2000; Chiarella et al., 2009). Brock and Hommes (1998) adopt a simple linear representation of beliefs:
where is the trend component and the bias of trader type . If , the agent can be either a pure trend chaser if (strong trend chaser if ), or a contrarian if (strong contrarian if ). If , the agent of type is purely biased (upward or downward biased if or . In the special case when both and are equal to zero, the agent is a “fundamentalists”, i.e. she believes that prices return to their fundamental value. Agents can also be fully rational, with . In such a case, they have perfect foresight but, they must pay a cost .121212In our experiments we allow for the possibility that a positive cost might be by paid also by non-rational traders. This mirrors the fact that some trader might want to buy additional information, which they might not be able to use (due e.g. to computational mistakes).
In our application, we use a simple model with only two types of agents, whose behaviours vary according to the choice of trend components, biases and perfect forecasting costs. Combining equations (7), (9) and (10), one can derive the following equilibrium condition:
which allows to compute the price of the risky asset (in deviation from the fundamental) at time .
Traders switch among different strategies according to the their evolving profitability. More specifically, each strategy is associated with a fitness measure of the form:
where is a weight attributed to past profits. At the beginning of each period, agents reassess the profitability of their trading strategy with respect to the others. The probability that an agent choose strategy is given by:
where the parameter captures traders’ intensity of choice. According to equation 13, successful strategies gain an increasing number of followers. In addition, the algorithm introduces a certain amount of randomness, as less profitable strategies may still be chosen by traders. In this way, the model captures imperfect information and agents’ bounded rationality. Moreover, the system can never be stacked in an equilibrium where all traders adopt the same strategy.
4.2 Experimental design and empirical setting
Despite the model is relatively simple, different contributions have tried to match the statistical properties of its output with those observed in real financial markets (Boswijk et al., 2007; Recchioni et al., 2015; Lamperti, 2016; Kukacka and Barunik, 2016). This makes the model an ideal test case for our surrogate: it is relatively cheap in terms of computational needs, it offers a reasonably large parameter space and it has been extensively studied in the literature.
There are 12 free parameters (Table 1) determined through calibration.131313We underline that the dimension of the parameter space is in line or even larger that in recent studies on ABM meta-modelling (see e.g. Salle and Yildizoglu, 2014; Bargigli et al., 2016). The ranges for parameters’ values have been identified relying on both economic reasoning and previous experiments on the model. However, their selection is ultimately a user specific decision. Our procedure manages the computational constraints faced by modellers working with large parameter spaces. In what follows, we refer to the parameter space spanned by the intervals specified in the last column of Table 1. Naturally, it can be further expanded or reduced according to the user’s needs and the available budget.
Let us now consider the conditions identifying positive calibrations. As already discussed above, any feature of model’s output can be employed to express such conditions. According to Section 3 two types of calibration criteria are considered, giving respectively binary and real-valued outcomes. In the binary outcome case, we employ a two samples Kolmogorov-Smirnov (KS) test between the distribution of logarithmic returns obtained from the numerical simulation of the model and the one obtained from real stock market data.141414Let and be the prices of an asset at two subsequent time steps. The logarithmic return from to is given by . More specifically, we rely on daily adjusted closing prices for the S&P 500 from December 09, 2013 to December 07, 2015, for a total of 502 observations, and we compute the following test statistic:151515The data have been obtained from Yahoo Finance: https://finance.yahoo.com/quote/%5EGSPC/history. The test is passed if the null hypothesis “equality of the distributions” is not rejected at 5% confidence level.
where indicates logarithmic returns and and are the empirical distribution functions of the real world () and simulated () samples respectively. Then, in a real-valued outcome setting, we use the p-value of the KS test, , as an expression of model’s fit with the data. We also consider an equivalent condition for binary outcomes, where predicted labels with a p-value above are considered as positive calibrations. This choice is intentional as equivalent conditions allow a comparison between the binary and real-valued outcomes in terms of precision (ability to identify true calibrations) and computational time (in the real-valued scenario there is more information to be processed.)
We train the surrogate 100 times over 10 different budgets of 250, 500, 750, 1000, 1250, 1500, 1750, 2000, 2250, 2500 labelled parameter combinations and evaluate it on 100000 unlabelled points. Having a large number of out-of-sample, unlabelled, possibly well-spread points is fundamental to evaluate the performance of the meta-model. We use a larger evaluation set than any other meta-modelling contribution we are aware of (see, for instance, Salle and Yildizoglu, 2014; Dosi et al., 2017c; Bargigli et al., 2016).
In Figure 4, we show the parameter importance results for the Brock and Hommes (B&H) model. We find that the most relevant parameters to fit the empirical distribution of returns observed in the SP500 are those characterizing traders’ attitude towards the trend ( and ) and, secondly, their bias ( and ). This result is in line with recent findings by Recchioni et al. (2015) and Lamperti (2016) obtained using the same model. Moreover, the “intensity of choice” parameter (, cf. Section 4) is of crucial importance in the original model developed by Brock and Hommes (1998), but does not appear to be particularly relevant in determining the fit of the model with the data when compared to other behavioural parameters (at least within the range expressed by Table 1)161616See also Boswijk et al. (2007) where the authors estimate the B&H model on the SP500 and, in many exercises, find the switching parameter not to be significant.. Also traders’ risk attitude () and the weight associated to past profits () are relatively unimportant to shape the empirical performance of the model.
Let us now consider the behaviour of the surrogate. As outlined in Section 3.2, we run a series of exercises where the surrogate is employed to explore the behaviour of the model over the parameter space and filter out positive calibrations matching the distribution of real stock-market returns. Figure 5 collects the results and show the performance of the surrogate in the two proposed settings (binary and real-valued outcome). Within the binary outcome exercise, the -score increases with the size of the training sample and reaches a peak of around 0.80 when 2500 points are employed (cf. Figure 4(a)). In other words, the average between the share of true positive calibrations and the share of positive calibrations the surrogate correctly predicts is 0.8. Taking into consideration the upper bound of 1 and various practical applications (e.g. Petrovic et al., 2011; Cireşan et al., 2013), we consider the former result satisfying. However, such a classification performance should be evaluated in view of the surrogate’s searching ability, which is reported in Figure 4(c) and indicates the share of total positive calibration that the surrogate is able to find. Specifically, we find that around 75% of the positive calibrations present in the large set of out-of-sample points are found.
Obviously, the surrogate’s performance worsens as the training sample size is reduced. However, once we move to the real-valued setting, where the surrogate is learned using a continuous variable (containing more information about model’s behaviour), its performance is remarkably higher. Indeed, even when the sample size of the training points is particularly low (500), the True Positive Ratio (TPR) is around 70%, and reaches almost 95% (on average) when 2500 parameter vectors are employed (see Figure 4(d)).
Timing results are reported according to the average number of seconds required for a single compute core to complete the specific task 100 times. The increase in performance from classification (see Figure 4(e)) to regression (see Figure 4(f)) requires roughly 3X the modelling time and a nearly equivalent prediction time. Given this negligible prediction time, our approach facilitates a nearly costless exploration of the parameter space, delivering good results in terms of -score, TPR and MSE. The time savings in comparison to running the original ABM are substantial. In this exercise over a set of 10000 out-of-sample points, the surrogate is 500X faster on average in prediction. Note also that the learned surrogate is reusable on any number of out-of-sample parameter combinations, without the need for additional training. Further, we remark that computational gains are expected to be larger as more complex and expensive-to-simulate models are used. The next section goes in this direction.
5 Surrogate modelling examples: the Islands model
In the “Island” growth model (Fagiolo and Dosi, 2003), a population of heterogeneous firms locally interact discovering and diffusing new technologies, which ultimately lead to the emergence (or not) of endogenous growth. After presenting the model in Section 5.1, we describe the empirical setting (see Section 5.2) and the results of the machine learning calibration and exploration exercises (cf. Section 5.3). Recall that the seed used for the pseudo-random number generator is fixed and kept constant across runs of the model over different parameter vectors.
5.1 The Island growth model
A fixed population of heterogeneous firms () explore an unknown technological space (“the sea”), punctuated by islands (indexed by ) representing new technologies. The technological space is represented by a 2-dimensional, infinite, regular lattice endowed with the Manhattan metrics . The probability that each node is an island is equal to . There is only one homogeneous good, which can be “mined” from any island. Each island is characterized by a productivity coefficient . The production of agent on island having coordinates is equal to:
where and indicates the total number of miners working on at time . The GDP of the economy is simply obtained by summing up the production of each island.
Each agent can choose to be a miner and produce an homogeneous final good in her current island, to become an explorer and search for new islands (i.e. technologies), or to be an imitator and moves towards a known island. In each time step, miners can decide to become explorer with probability . In that case, the agent leaves the island and “sails” around until another (possibly still unknown) island is discovered. During the search, explorers are not able to extract any output and randomly move in the lattice. When a new island (technology) is discover, its productivity is given by:
where is a Poisson distributed random variable with mean , is a uniformly distributed random variable with zero mean and unitary variance, is a constant between zero and one and, finally, is the output memory of agent . Therefore, the initial productivity of a newly discovered island depends on four factors (see Dosi, 1988): (i) its distance from the origin; (ii) cumulative learning effects (); (iii) a random variable capturing radical innovations (i.e. changes in technological paradigms); (iv) a stochastic i.i.d. zero-mean noise controlling for high-probability low-jumps (i.e. incremental innovations).
Miners can also decide to imitate currently available technologies by taking advantage of informational spill-overs stemming from more productive islands located in their technological neighbourhoods. More specifically, agents mining on any colonized island deliver a signal, which is instantaneously spread in the system. Other agents in the lattice receive the signal with probability:
which depends on the magnitude of technology gap as well as on the physical distance between two islands (). Agent chooses the strongest signal and become an imitator sealing to island according to the shortest possible path. Once the imitated island is reached, the imitator will start mining again.
The model shows that the very possibility of notionally unlimited (albeit unpredictable) technological opportunities is a necessary condition for the emergence of endogenous exponential growth. Indeed, self-sustained growth is achieved whenever technological opportunities (captured by both the density of islands and the likelihood of radical innovations ), path-dependency (i.e. the fraction of idiosyncratic knowledge, , agents carry over to newly discovered technologies), and spreading intensity in the information diffusion process (), are beyond some minimum thresholds (Fagiolo and Dosi, 2003). Moreover, the system endogenously generate exponential growth if the trade-off between exploration and exploitation is solved, i.e. if the ecology of agents find the right balance between searching for new technologies and mastering the available ones.
5.2 Experimental design and empirical setting
The Island model employs eight input parameters to generate a wide array of growth dynamics. We report the parameters, their theoretical support and the explored range in Table 2. We kept the number of firms fixed (and equal to 50) to study what happens to the same economic system, when the parameters linked to behavioural rules are changed.171717Note that the Island model does not exhibit scale effects: the results generated by the model does not depend on the number of agents in the system (Fagiolo and Dosi, 2003).
Similarly to section 4.2, we characterize a binary outcome and a real-valued outcome setting. In the first case, the surrogate is learnt using a binary target variable taking value 1 if a user-defined specific set of conditions is satisfied and zero otherwise. More specifically, we define two conditions characterizing the GDP time series generated by the model. The first condition requires the model to generate self-sustained sustained pattern of output growth. Given the long-run average growth rate of the economy ():
sustained growth emerges if .
The second condition aims at capturing the presence of fat tails in the output growth-rate distributions. These empirical regularities suggest that deep downturns can coexist with mild fluctuations and has been found in both OECD (Fagiolo et al., 2008) and developing countries (Castaldi and Dosi, 2009; Lamperti and Mattei, 2016). More specifically, we fit a symmetric exponential power distribution (see Subbotin, 1923; Bottazzi and Secchi, 2006) , whose functional form reads:
where controls for the standard deviation, for the shape of the distribution and represents the mean. As gets smaller, the tails become fatter. In particular, when the distribution reduces to a Gaussian one, while for the density is Laplacian. We say that the output growth-rate distribution exhibits fat tails if . Note that there is a hierarchy in the conditions we have just defined: only those parametrizations satisfying the first one () are retained as candidates for positive calibrations and further investigated with respect to the second condition. In the real-valued outcome case, instead, we just focus on shape of growth rates distribution. In particular, we our target variable is the estimated of the symmetric power exponential distribution and a positive calibration is found if .181818In the real-valued outcome setting our exercise is comparable to those performed in Dosi et al. (2017c), where the same distribution and parameters are used in a model of industrial dynamics. Again, the choice of the condition to be satisfied ensures (partial, in this case) consistency between the two settings.
We train the surrogate as we did with the B&H model, but given the higher computational complexity of the Island model, we reduce the number of unlabelled points to 10000.191919This choice is motivated by the fact that we need to run the model on the out-of-sample points in order to evaluate the surrogate.
As for the Brock and Hommes model, we start our analysis reporting the relative importance for all the parameters characterizing the Island model (figure 6). We find that all the parameters of the model linked to production, innovation and imitation appear to be relevant for the emergence of sustained economic growth.
The surrogate’s performances is presented in Figure 7, where the first column of the plots refers to the binary outcome setting, while the second one to the real-valued one. The -score displays relatively high values even for low training sample sizes (250 and 500) pointing to a good classification performance of the surrogate (see Figure 6(a)). However, it quickly saturates, reaching a plateau around 0.8. Conversely, in the real-valued setting, the surrogate’s performance keeps increasing with the training sample size, and it displays remarkably low values of MSE when more than 1000 points are employed (cf. Figure 6(b)).
In both settings, the searching ability of the surrogate behaves in a similar way: the TRP steadily increases with the training sample size (cf. Figures 6(c) and 6(d)). In absolute terms, the real-valued setting delivers much better results than the binary one, as for the Brock and Hommes model (section 4.3). In particular, the largest true positive ratio reaches 0.9 for the real-valued case and 0.8 for the binary one. Therefore, by training the surrogate on 2500 points we are able to (i) find 90% of true positive calibrations (Figure 6(d)) and predict the thickness of the associated distribution of growth rates incurring in a mean squared error of less than 0.08 (Figure 6(b)) using a continuous target variable and, (ii) find 80% of the true positives (panel 6(c)) and correctly classifying around the 80% of them (panel 6(a)) using a binary target variable.
Given the satisfactory explanatory performance of the surrogate, do we also achieve considerable improvements in the computational time required to perform such exploration exercises? Figures 6(e) and 6(f) provides a positive answer. Indeed, the surrogate is 3750 times faster than the fully-fledge Island agent-based model. Moreover, the increase in speed is considerably larger than in the Brock and Hommes model. This confirms our intuition on the increasing usefulness of our surrogate modeling approach when the computational cost of the ABM under study is higher. Such a result is a desirable property for real applications, where the complexity of the underlying ABM could even prevent the exploration of the parameter space.
5.4 Robustness analysis
We now assess the robustness of our training procedure with respect to different surrogate models. More specifically, we compare the XGBoost surrogate employed in the previous analysis with the simpler and more widely used Logit one. Our comparison exercise is performed in a fully stochastic version of the Island agent-based model, where an additional Monte Carlo (MC) is carried out on the seed parameter governing the stochastic terms of the model.
We focus on a binary outcome setting (the one delivering worse performances) and we employ the milder condition that the average growth rate must be positive and sustained, i.e. . In this way, the results can be compared to those obtained in the original exercise in Fagiolo and Dosi (2003). We set a budget of 500 evaluations of the “true” Islands ABM and run a Monte Carlo exercise of size 100 per parameter combination to generate an MC average of the GDP growth rate that serves as our output variable. Note that this exercise is more complete that the one performed in the previous sections: here, we develop a surrogate model that learns the relationship between parameters and the MC average over their ABM evaluations. Note that this requires many more evaluations of the parameter combination in the true ABM to converge to the statistic required for the label. In our proposed procedure, an MC average growth rate below 0.5% is labelled “false”, while AGR above 0.5% are labeled “true”. The aim is to learn a surrogate model that accurately classifies parameter combinations as positive or negative calibrations.
We demonstrate the performance of our active learning approach using two different surrogates: the XGBoost and the faster, less precise, Logit. The former, employed in the analyses carried out in the previous sections, benefits from increased accuracy in exchange for greater computational costs. The latter is a standard statistical model employed regularly for this type of regression analysis. The performance of these alternative surrogates will be evaluated according to the F1-score while training the surrogate, with the final objective of maximizing the precision of the resulting models, i.e. the number of true evaluations which are accurately predicted as positive before they are evaluated. This is a key point to this exercise because real-world use of the proposed approach does not allow us to evaluate all the points in our sample space. Real-world evaluation only provides labels for points that are predicted positive and the resulting performance can only be measured with regard to the true and false positives, with a preference to maximize the former.
The exercise is performed using the Python BOASM package.202020See https://github.com/amirsani/BOASM The algorithm mirror exactly the one described in Section 3. The exercise begins by sampling points at random from the Islands parameter space. Given the fixed budget of evaluations of the true ABM, for both the XGBoost and Logit, the first surrogate is provided with labelled parameters selected at random from the points, according to the total-variation sampling procedure in Saltelli et al. (2010). Then, over several rounds, a surrogate will be fit to the labelled parameters and used to predict a labelling over the points. The predicted labels will then be employed by the proposed procedure to select points that will be added at each round to the set of labelled points. A new surrogate is learned in the subsequent round and the procedure will repeat until the budget of true evaluations has been reached.
|Surrogate Algorithm||True Negatives||False Positives||False Negatives||True Positives||Precision|
The proposed procedure results in a comparable precision of 94.17% and 94.72% between Logit and XGBoost, respectively. The negligible difference between the precision of the two surrogates suggests that out training procedure provides satisfying results even when the standard Logit statistical model is employed. However, when the XGBoost predicted probabilities are corrected through the Platt scaling procedure,212121Unlike Logit, which produces accurate probabilities for each of the class labels, probabilities produced by non-parametric algorithms such as XGBoost require scaling. Here, we use Platt Scaling to correct the probabilities produced with XGBoost. For more information, see Platt et al. (1999). its precision rises to 99.35%. Moreover, scaled XGBoost performs is considerably superior to Logit with regard to true vs. false positives. Considering its higher computation costs and need for hyperparameter optimization in using the more precise XGBoost surrogate, users might prefer the faster Logit surrogate when false positives are cheap. Nevertheless, our proposed surrogate modelling procedure works well in both the Logit and XGBoost cases.
6 Discussion and concluding remarks
In this paper, we have proposed a novel approach to the calibration and parameter space exploration of agent-based models (ABM), which combines the use of supervised machine learning and intelligent sampling to construct a cheap surrogate meta-model. To the best of our knowledge, this is the first attempt to exploit machine-learning techniques for calibration and exploration in an agent-based framework.
Our machine-learning surrogate approach is different from kriging, which has been recently applied to ABMs dealing with industrial dynamics (Salle and Yildizoglu, 2014; Dosi et al., 2017c), financial networks (Bargigli et al., 2016) and macroeconomic issues (Dosi et al., 2016, 2017b). In particular, apart from the different statistical framework kriging relies on (it assumes a multivariate Gaussian process), the results it delivers once applied to ABMs may suffer from three relevant limitations. First, kriging is difficult to apply to large scale models, where the number of parameters goes beyond 20. This constrains the modeller to introduce additional procedures to select, a priori, the subset of parameters to study, while leaving the rest constant (see e.g. Dosi et al., 2016). Second, the machine-learning surrogate approach performs better in out-of-sample testing: the typical kriging-based meta-model is tested on 10-20 points within an extremely large space, while our surrogate is tested on samples with size 10000 in the first set of exercises and 1000000 points in the last exercise. Finally, the response surfaces generated by kriging meta-models suffer from smoothness assumptions that collapse interesting patterns, which cannot be captured by common Gaussianity assumptions. This results in incredibly smooth and well-behaved surfaces, which may falsely relate parameters and model behaviour. Given the ragged, unsmooth surfaces commonly reported in agent-based models (see e.g. Gilli and Winker, 2003; Fabretti, 2012; Lamperti, 2016), inferring the behaviour of the true ABM on the basis of the insights produced by kriging meta-model may results in large errors. Further, even when smooth response surfaces exist, kriging requires the selection of the correct prior and kernel. Even in state-of-the-art likelihood-free approaches, one must select the correct sufficient statistic and acceptance threshold to provide any value.
The proposed approach manages all these problems.222222In the current work, we also focus on examples dealing with relatively few parameters. This choice is motivated by illustrative reasons and the willingness to use well established models whose code is easily replicable. Further, the results from this paper were produced using a relatively common laptop computer with 16 gigabytes of memory and a 2.4Ghz Intel i7 5500 CPU. The application to a large scale model is currently under development. However, the computational parsimony of the algorithm used to construct our surrogates strongly points to the ability to deal with much richer parameter spaces. However, the main advantage of our methodology remains in its practical usefulness. Indeed, the surrogate can be learnt at virtually zero computational costs (for research applications) and requires a trivial amount of time to predict areas of the parameter space the modeller should focus on with reasonably good results. Two modelling options are presented, a binary outcome setting and a real-valued one: the first is faster and especially useful when a large number of samples is available, while the second has more explanatory power. Furthermore, the usual trade-off between the quantity of information that needs to be processed (computational costs) and the surrogate performance improvements is, in practice, absent. Ultimately, the surrogate prediction exercises proposed in this paper take less than a minute to complete, with the majority of computation coming from the time to assess the budget of true ABM model evaluations. This means, in practical terms, that the modeller can use an arbitrarily large set of parameter combinations and a relatively small training sample to build the surrogate at almost no cost and leverage the resulting meta-model to gain an insight on the dynamics of the parameter space for further exploration using the original ABM.
Finally, an additional relevant result emerges from the exercises investigated in this paper. The surrogate is much more effective in reducing the relative cost of exploring the properties of the model over the parameter space for the “Islands” model, which is more computationally intensive than the Brock and Hommes. This suggests that the adoption of surrogate meta-modelling allows to achieve increasing computational gains as the complexity of the underlying model increases.
This work is only the first step towards a comprehensive assessment of agent-based model properties through machine-learning techniques. Such developments are especially important for complex macroeconomic agent-based models (see e.g. Dosi et al., 2010, 2013, 2015, 2017a; Popoyan et al., 2017) as they could allow the development of a standardized and robust procedure for model calibration and validation, thus closing the existing gap with Dynamics Stochastic General Equilibrium models (see Fagiolo and Roventini, 2017, for a critical comparison of ABM and DSGE models). Accordingly, a user-friendly Python surrogate modelling library will also be released for general use.
A special thank goes to Antoine Mandel, who constantly engaged in fruitful discussions with the authors and provided incredibly valuable insights and suggestions. We would also like to thank Daniele Giachini, Mattia Guerini, Matteo Sostero and Baláz Kégl for their comments. Further, we would like to thank all the participants in seminars and workshops held at Scuola Superiore Sant’Anna (Pisa), the XXI WEHIA conference (Castellon), the XXII CEF conference (Bordeaux), NIPS 2016 ”What If? Inference and Learning of Hypothetical and Counterfactual Interventions in Complex Systems” workshop (Barcelona), CCS 2016 conference (Amsterdam) and 2016 Paris-Bielefeld Workshop on Agent-Based Modeling (Paris). FL acknowledges financial support from European Union’s FP7 project IMPRESSIONS (G.A. No 603416). AS acknowledges financial support from the H2020 project DOLFINS (G.A. No 640772) and hardware support from NVIDIA Corporation, AdapData SAS and the Grid5000 testbed for this research. AR acknowledges financial support from European Unionâs FP7 IMPRESSIONS, H2020 DOLFINS and H2020 ISIGROWTH (G.A. No 649186) projects.
- Alfarano et al. (2005) Alfarano, S., Lux, T., and Wagner, F. (2005). Estimation of agent-based models: The case of an asymmetric herding model. Computational Economics, 26(1):19–49.
- Alfarano et al. (2006) Alfarano, S., Lux, T., and Wagner, F. (2006). Estimation of a simple agent-based model of financial markets: An application to australian stock and foreign exchange data. Physica A: Statistical Mechanics and its Applications, 370(1):38 – 42.
- Amilon (2008) Amilon, H. (2008). Estimation of an adaptive stock market model with heterogeneous agents. Journal of Empirical Finance, 15(2):342 – 362.
- An and Wilensky (2009) An, G. and Wilensky, U. (2009). From artificial life to in silico medicine. In Komosinski, M. and Adamatzky, A., editors, Artificial Life Models in Software, pages 183–214. Springer London, London.
- Anderson et al. (1972) Anderson, P. W. et al. (1972). More is different. Science, 177(4047):393–396.
- Archer and Kimes (2008) Archer, K. J. and Kimes, R. V. (2008). Empirical characterization of random forest variable importance measures. Computational Statistics & Data Analysis, 52(4):2249–2260.
- Assenza et al. (2015) Assenza, T., Gatti, D. D., and Grazzini, J. (2015). Emergent dynamics of a macroeconomic agent based model with capital and credit. Journal of Economic Dynamics and Control, 50:5–28.
- Banerjee (1992) Banerjee, A. V. (1992). A simple model of herd behavior. The Quarterly Journal of Economics, 107(3):797–817.
- Barde (2016a) Barde, S. (2016a). Direct comparison of agent-based models of herding in financial markets. Journal of Economic Dynamics and Control, 73:329 – 353.
- Barde (2016b) Barde, S. (2016b). A practical, accurate, information criterion for nth order markov processes. Computational Economics, pages 1–44.
- Bargigli et al. (2016) Bargigli, L., Riccetti, L., Russo, A., and Gallegati, M. (2016). Network Calibration and Metamodeling of a Financial Accelerator Agent Based Model. Working papers, economics, Universitá degli Studi di Firenze, Dipartimento di Scienze per l’Economia e l’Impresa.
- Bergstra and Bengio (2012) Bergstra, J. and Bengio, Y. (2012). Random search for hyper-parameter optimization. Journal of Machine Learning Research, 13(Feb):281–305.
- Booker et al. (1999) Booker, A., Dennis, J.E., J., Frank, P., Serafini, D., Torczon, V., and Trosset, M. (1999). A rigorous framework for optimization of expensive functions by surrogates. Structural optimization, 17(1):1–13.
- Boswijk et al. (2007) Boswijk, H., Hommes, C., and Manzan, S. (2007). Behavioral heterogeneity in stock prices. Journal of Economic Dynamics and Control, 31(6):1938 – 1970.
- Bottazzi and Secchi (2006) Bottazzi, G. and Secchi, A. (2006). Explaining the distribution of firm growth rates. The RAND Journal of Economics, 37(2):235–256.
- Breiman (2001) Breiman, L. (2001). Random forests. Machine learning, 45(1):5–32.
- Breiman et al. (1984) Breiman, L., Friedman, J., Stone, C. J., and Olshen, R. A. (1984). Classification and regression trees. CRC press.
- Brock and Hommes (1997) Brock, W. A. and Hommes, C. H. (1997). A rational route to randomness. Econometrica, 65(5):1059–1095.
- Brock and Hommes (1998) Brock, W. A. and Hommes, C. H. (1998). Heterogeneous beliefs and routes to chaos in a simple asset pricing model. Journal of Economic Dynamics and Control, 22(8–9):1235 – 1274.
- Brown et al. (2005) Brown, D. G., Page, S., Riolo, R., Zellner, M., and Rand, W. (2005). Path dependence and the validation of agent-based spatial models of land use. International Journal of Geographical Information Science, 19(2):153–174.
- Caiani et al. (2016) Caiani, A., Godin, A., Caverzasi, E., Gallegati, M., Kinsella, S., and Stiglitz, J. E. (2016). Agent based-stock flow consistent macroeconomics: Towards a benchmark model. Journal of Economic Dynamics and Control, 69:375–408.
- Carley et al. (2006) Carley, K. M., Fridsma, D. B., Casman, E., Yahja, A., Altman, N., Chen, L.-C., Kaminsky, B., and Nave, D. (2006). Biowar: scalable agent-based model of bioattacks. IEEE Transactions on Systems, Man, and Cybernetics-Part A: Systems and Humans, 36(2):252–265.
- Castaldi and Dosi (2009) Castaldi, C. and Dosi, G. (2009). The patterns of output growth of firms and countries: Scale invariances and scale specificities. Empirical Economics, 37(3):475–495.
- Chen et al. (2012) Chen, S.-H., Chang, C.-L., and Du, Y.-R. (2012). Agent-based economic models and econometrics. The Knowledge Engineering Review, 27:187–219.
- Chen and Guestrin (2016) Chen, T. and Guestrin, C. (2016). Xgboost: A scalable tree boosting system. In Proceedings of the 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 785–794. ACM.
- Chiarella et al. (2009) Chiarella, C., Iori, G., and Perelló, J. (2009). The impact of heterogeneous trading rules on the limit order book and order flows. Journal of Economic Dynamics and Control, 33(3):525–537.
- Cireşan et al. (2013) Cireşan, D. C., Giusti, A., Gambardella, L. M., and Schmidhuber, J. (2013). Mitosis detection in breast cancer histology images with deep neural networks. In International Conference on Medical Image Computing and Computer-assisted Intervention, pages 411–418. Springer.
- Claesen et al. (2014) Claesen, M., Simm, J., Popovic, D., Moreau, Y., and De Moor, B. (2014). Easy hyperparameter search using optunity. arXiv preprint arXiv:1412.1114.
- Conti and O’Hagan (2010) Conti, S. and O’Hagan, A. (2010). Bayesian emulation of complex multi-output and dynamic computer models. Journal of statistical planning and inference, 140(3):640–651.
- Dawid et al. (2014a) Dawid, H., Gemkow, S., Harting, P., Van der Hoog, S., and Neugart, M. (2014a). Agent-based macroeconomic modeling and policy analysis: the eurace@ unibi model. Technical report, Bielefeld Working Papers in Economics and Management.
- Dawid et al. (2014b) Dawid, H., Harting, P., and Neugart, M. (2014b). Economic convergence: Policy implications from a heterogeneous agent model. Journal of Economic Dynamics and Control, 44:54–80.
- De Marchi (2005) De Marchi, S. (2005). Computational and mathematical modeling in the social sciences. Cambridge University Press.
- Dosi (1988) Dosi, G. (1988). Sources, procedures and microeconomic effects of innovation. Journal of Economic Literature, 26:126–71.
- Dosi et al. (2013) Dosi, G., Fagiolo, G., Napoletano, M., and Roventini, A. (2013). Income distribution, credit and fiscal policies in an agent-based Keynesian model. Journal of Economic Dynamics and Control, 37(8):1598–1625.
- Dosi et al. (2015) Dosi, G., Fagiolo, G., Napoletano, M., Roventini, A., and Treibich, T. (2015). Fiscal and monetary policies in complex evolving economies. Journal of Economic Dynamics and Control, 52(C):166–189.
- Dosi et al. (2010) Dosi, G., Fagiolo, G., and Roventini, A. (2010). Schumpeter meeting Keynes: A policy-friendly model of endogenous growth and business cycles. Journal of Economic Dynamics and Control, 34(9):1748–1767.
- Dosi et al. (2017a) Dosi, G., Pereira, M., Roventini, A., and Virgillito, M. (2017a). When more flexibility yields more fragility: The microfoundations of keynesian aggregate unemployment. Journal of Economic Dynamics and Control, forthcoming.
- Dosi et al. (2016) Dosi, G., Pereira, M., Roventini, A., and Virgillito, M. E. (2016). The Effects of Labour Market Reforms upon Unemployment and Income Inequalities: an Agent Based Model. LEM Working Papers Series 2016-27, Scuola Superiore Sant’Anna.
- Dosi et al. (2017b) Dosi, G., Pereira, M., Roventini, A., and Virgillito, M. E. (2017b). Causes and consequences of hysteresis: Aggregate demand, productivity and employment. LEM Working Papers Series 2017-07, Scuola Superiore Sant’Anna.
- Dosi et al. (2017c) Dosi, G., Pereira, M. C., and Virgillito, M. E. (2017c). On the robustness of the fat-tailed distribution of firm growth rates: a global sensitivity analysis. Journal of Economic Interaction and Coordination, pages 1–21.
- Effken et al. (2012) Effken, J. A., Carley, K. M., Lee, J.-S., Brewer, B. B., and Verran, J. A. (2012). Simulating nursing unit performance with orgahead: strengths and challenges. Computers, informatics, nursing: CIN, 30(11):620.
- Fabretti (2012) Fabretti, A. (2012). On the problem of calibrating an agent based model for financial markets. Journal of Economic Interaction and Coordination, 8(2):277–293.
- Fagiolo et al. (2007) Fagiolo, G., Birchenhall, C., and Windrum, P. (2007). Empirical validation in agent-based models: Introduction to the special issue. Computational Economics, 30(3):189–194.
- Fagiolo and Dosi (2003) Fagiolo, G. and Dosi, G. (2003). Exploitation, exploration and innovation in a model of endogenous growth with locally interacting agents. Structural Change and Economic Dynamics, 14(3):237–273.
- Fagiolo et al. (2008) Fagiolo, G., Napoletano, M., and Roventini, A. (2008). Are output growth-rate distributions fat-tailed? some evidence from oecd countries. Journal of Applied Econometrics, 23(5):639–669.
- Fagiolo and Roventini (2012) Fagiolo, G. and Roventini, A. (2012). Macroeconomic policy in dsge and agent-based models. Revue de l’OFCE, 124:67–116.
- Fagiolo and Roventini (2017) Fagiolo, G. and Roventini, A. (2017). Macroeconomic policy in dsge and agent-based models redux: New developments and challenges ahead. Journal of Artificial Societies and Social Simulation, 20(1).
- Feurer et al. (2015) Feurer, M., Klein, A., Eggensperger, K., Springenberg, J., Blum, M., and Hutter, F. (2015). Efficient and robust automated machine learning. In Advances in Neural Information Processing Systems, pages 2962–2970.
- Franke (2009) Franke, R. (2009). Applying the method of simulated moments to estimate a small agent-based asset pricing model. Journal of Empirical Finance, 16(5):804 – 815.
- Franke and Westerhoff (2012) Franke, R. and Westerhoff, F. (2012). Structural stochastic volatility in asset pricing dynamics: Estimation and model contest. Journal of Economic Dynamics and Control, 36(8):1193–1211.
- Freund (1990) Freund, Y. (1990). Boosting a weak learning algorithm by majority. In COLT, volume 90, pages 202–216.
- Freund et al. (1996) Freund, Y., Schapire, R. E., et al. (1996). Experiments with a new boosting algorithm. In Icml, volume 96, pages 148–156.
- Gallegati and Kirman (2012) Gallegati, M. and Kirman, A. (2012). Reconstructing economics. Complexity Economics, 1(1):5–31.
- Gilli and Winker (2003) Gilli, M. and Winker, P. (2003). A global optimization heuristic for estimating agent based models. Computational Statistics & Data Analysis, 42(3):299 – 312. Computational Ecomometrics.
- Goldberg et al. (2011) Goldberg, A. B., Zhu, X., Furger, A., and Xu, J.-M. (2011). Oasis: Online active semi-supervised learning. In AAAI.
- Grazzini (2012) Grazzini, J. (2012). Analysis of the emergent properties: Stationarity and ergodicity. Journal of Artificial Societies and Social Simulation, 15(2):7.
- Grazzini and Richiardi (2015) Grazzini, J. and Richiardi, M. (2015). Estimation of ergodic agent-based models by simulated minimum distance. Journal of Economic Dynamics and Control, 51:148 – 165.
- Grazzini et al. (2017) Grazzini, J., Richiardi, M. G., and Tsionas, M. (2017). Bayesian estimation of agent-based models. Journal of Economic Dynamics and Control, 77:26 – 47.
- Grimm and Railsback (2013) Grimm, V. and Railsback, S. F. (2013). Individual-based modeling and ecology. Princeton university press.
- Guerini and Moneta (2016) Guerini, M. and Moneta, A. (2016). A Method for Agent-Based Models Validation. LEM Papers Series 2016/16, Laboratory of Economics and Management (LEM), Sant’Anna School of Advanced Studies, Pisa, Italy.
- Herlands et al. (2015) Herlands, W., Wilson, A., Nickisch, H., Flaxman, S., Neill, D., Van Panhuis, W., and Xing, E. (2015). Scalable gaussian processes for characterizing multidimensional change surfaces. arXiv preprint arXiv:1511.04408.
- Ilachinski (1997) Ilachinski, A. (1997). Irreducible semi-autonomous adaptive combat (isaac): An artificial-life approach to land warfare. Technical report, DTIC Document.
- Kukacka and Barunik (2016) Kukacka, J. and Barunik, J. (2016). Estimation of financial agent-based models with simulated maximum likelihood. IES Working Paper 7/2016, Charles University of Prague.
- Lamperti (2016) Lamperti, F. (2016). Empirical Validation of Simulated Models through the GSL-div: an Illustrative Application. LEM Papers Series 2016/18, Laboratory of Economics and Management (LEM), Sant’Anna School of Advanced Studies, Pisa, Italy.
- Lamperti (2017) Lamperti, F. (2017). An information theoretic criterion for empirical validation of simulation models. Econometrics and Statistics, forthcoming.
- Lamperti et al. (2017) Lamperti, F., Dosi, G., Napoletano, M., Roventini, A., and Sapio, A. (2017). Faraway, so close: coupled climate and economic dynamics in an agent based integrated assessment model. Lem working papers series, Scuola Superiore Sant’Anna.
- Lamperti and Mattei (2016) Lamperti, F. and Mattei, C. E. (2016). Going Up and Down: Rethinking the Empirics of Growth in the Developing and Newly Industrialized World. LEM Papers Series 2016/01, Laboratory of Economics and Management (LEM), Sant’Anna School of Advanced Studies, Pisa, Italy.
- Leal et al. (2014) Leal, S. J., Napoletano, M., Roventini, A., and Fagiolo, G. (2014). Rock around the clock: an agent-based model of low-and high-frequency trading. Journal of Evolutionary Economics, pages 1–28.
- Lee et al. (2015) Lee, J.-S., Filatova, T., Ligmann-Zielinska, A., Hassani-Mahmooei, B., Stonedahl, F., Lorscheid, I., Voinov, A., Polhill, J. G., Sun, Z., and Parker, D. C. (2015). The complexities of agent-based modeling output analysis. Journal of Artificial Societies and Social Simulation, 18(4):4.
- Li et al. (2013) Li, X., Engelbrecht, A., and Epitropakis, M. G. (2013). Benchmark functions for cec’2013 special session and competition on niching methods for multimodal function optimization. RMIT University, Evolutionary Computation and Machine Learning Group, Australia, Tech. Rep.
- Louppe et al. (2013) Louppe, G., Wehenkel, L., Sutera, A., and Geurts, P. (2013). Understanding variable importances in forests of randomized trees. In Advances in neural information processing systems, pages 431–439.
- Lux and Marchesi (2000) Lux, T. and Marchesi, M. (2000). Volatility clustering in financial markets: a microsimulation of interacting agents. International journal of theoretical and applied finance, 3(04):675–702.
- Macy and Willer (2002) Macy, M. W. and Willer, R. (2002). From factors to actors: Computational sociology and agent-based modeling. Annual review of sociology, pages 143–166.
- Marks (2013) Marks, R. E. (2013). Validation and model selection: Three similarity measures compared. Complexity Economics, 2(1):41–61.
- Morokoff and Caflisch (1994) Morokoff, W. J. and Caflisch, R. E. (1994). Quasi-random sequences and their discrepancies. SIAM Journal on Scientific Computing, 15(6):1251–1279.
- Moss (2008) Moss, S. (2008). Alternative approaches to the empirical validation of agent-based models. Journal of Artificial Societies and Social Simulation, 11(1):5.
- Petrovic et al. (2011) Petrovic, S., Osborne, M., and Lavrenko, V. (2011). Rt to win! predicting message propagation in twitter. ICWSM, 11:586–589.
- Platt et al. (1999) Platt, J. et al. (1999). Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in large margin classifiers, 10(3):61–74.
- Popoyan et al. (2017) Popoyan, L., Napoletano, M., and Roventini, A. (2017). Taming Macroeconomic Instability: Monetary and Macro Prudential Policy Interactions in an Agent-Based Model. Journal of Economic Behavior & Organization, 134:117–140.
- Rasmussen and Williams (2006) Rasmussen, C. E. and Williams, C. K. I. (2006). Gaussian processes for machine learning. MIT Press.
- Recchioni et al. (2015) Recchioni, M. C., Tedeschi, G., and Gallegati, M. (2015). A calibration procedure for analyzing stock price dynamics in an agent-based framework. Journal of Economic Dynamics and Control, 60:1 – 25.
- Ross et al. (2011) Ross, S., Gordon, G. J., and Bagnell, D. (2011). A reduction of imitation learning and structured prediction to no-regret online learning. In AISTATS, volume 1(2), page 6.
- Ryabko (2016) Ryabko, D. (2016). Things bayes can’t do. In International Conference on Algorithmic Learning Theory, pages 253–260. Springer.
- Salle and Yildizoglu (2014) Salle, I. and Yildizoglu, M. (2014). Efficient Sampling and Meta-Modeling for Computational Economic Models. Computational Economics, 44(4):507–536.
- Saltelli et al. (2010) Saltelli, A., Annoni, P., Azzini, I., Campolongo, F., Ratto, M., and Tarantola, S. (2010). Variance based sensitivity analysis of model output. design and estimator for the total sensitivity index. Computer Physics Communications, 181(2):259–270.
- Settles (2010) Settles, B. (2010). Active learning literature survey. Technical Report 55-66, University of Wisconsin, Madison.
- Squazzoni (2010) Squazzoni, F. (2010). The impact of agent-based models in the social sciences after 15 years of incursions. History of Economic Ideas, pages 197–233.
- Subbotin (1923) Subbotin, M. T. (1923). On the law of frequency of error. Matematicheskii Sbornik, 31(2):296–301.
- ten Broeke et al. (2016) ten Broeke, G., van Voorn, G., and Ligtenberg, A. (2016). Which sensitivity analysis method should i use for my agent-based model? Journal of Artificial Societies & Social Simulation, 19(1).
- Tesfatsion and Judd (2006) Tesfatsion, L. and Judd, K. L. (2006). Handbook of computational economics: agent-based computational economics, volume 2. Elsevier.
- Thiele et al. (2014) Thiele, J. C., Kurth, W., and Grimm, V. (2014). Facilitating parameter estimation and sensitivity analysis of agent-based models: A cookbook using netlogo and r. Journal of Artificial Societies and Social Simulation, 17(3):11.
- van der Hoog (2016) van der Hoog, S. (2016). Deep Learning in Agent-Based Models: A Prospectus. Technical report, Faculty of Business Administration and Economics, Bielefeld University.
- Van Rijsbergen (1979) Van Rijsbergen, C. (1979). Information Retrieval. London: Butterworths.
- Weeks (1995) Weeks, M. (1995). Circumventing the curse of dimensionality in applied work using computer intensive methods. The Economic Journal, 105(429):520–530.
- Wilson et al. (2015) Wilson, A. G., Dann, C., and Nickisch, H. (2015). Thoughts on massively scalable gaussian processes. arXiv preprint arXiv:1511.01870.
- Winker et al. (2007) Winker, P., Gilli, M., and Jeleskovic, V. (2007). An objective function for simulation based inference on exchange rate data. Journal of Economic Interaction and Coordination, 2(2):125–145.
- Wolpert (2002) Wolpert, D. H. (2002). The supervised learning no-free-lunch theorems. In Soft Computing and Industry, pages 25–42. Springer.
- Wong (2015) Wong, K.-C. (2015). Evolutionary multimodal optimization: A short survey. arXiv preprint arXiv:1508.00457.
- Zhu (2005) Zhu, X. (2005). Semi-supervised learning literature survey. Technical report, University of Wisconsin-Madison.