Part of a series on |
Economics |
---|
Dynamic stochastic general equilibrium modeling (abbreviated as DSGE, or DGE, or sometimes SDGE) is a macroeconomic method which is often employed by monetary and fiscal authorities for policy analysis, explaining historical time-series data, as well as future forecasting purposes. [1] DSGE econometric modelling applies general equilibrium theory and microeconomic principles in a tractable manner to postulate economic phenomena, such as economic growth and business cycles, as well as policy effects and market shocks.
As a practical matter, people often use the term "DSGE models" to refer to a particular class of classically quantitative econometric models of business cycles or economic growth called real business cycle (RBC) models. [2] DSGE models were initially proposed by Kydland & Prescott, [3] and Long & Plosser; [4] Charles Plosser described RBC models as a precursor for DSGE modeling. [5]
As mentioned in the Introduction, DSGE models are the predominant framework of macroeconomic analysis. They are multifaceted, and their combination of micro-foundations and optimising economic behaviour of rational agents allows for a comprehensive analysis of macro effects. As indicated by their name, their defining characteristics are as follows:
The formulation and analysis of monetary policy has undergone significant evolution in recent decades and the development of DSGE models has played a key role in this process. As was aforementioned DSGE models are seen to be an update of RBC (real business cycle) models.
Early real business-cycle models postulated an economy populated by a representative consumer who operates in perfectly competitive markets. The only sources of uncertainty in these models are "shocks" in technology. [2] RBC theory builds on the neoclassical growth model, under the assumption of flexible prices, to study how real shocks to the economy might cause business cycle fluctuations. [7]
The "representative consumer" assumption can either be taken literally or reflect a Gorman aggregation of heterogenous consumers who are facing idiosyncratic income shocks and complete markets in all assets. [note 1] These models took the position that fluctuations in aggregate economic activity are actually an "efficient response" of the economy to exogenous shocks.
The models were criticized on a number of issues:
In a 1976 paper, [note 3] Robert Lucas argued that it is naive to try to predict the effects of a change in economic policy entirely on the basis of relationships observed in historical data, especially highly aggregated historical data. Lucas claimed that the decision rules of Keynesian models, such as the fiscal multiplier, cannot be considered as structural, in the sense that they cannot be invariant with respect to changes in government policy variables, stating:
This meant that, because the parameters of the models were not structural, i.e. not indifferent to policy, they would necessarily change whenever policy was changed. The so-called Lucas critique followed similar criticism undertaken earlier by Ragnar Frisch, in his critique of Jan Tinbergen's 1939 book Statistical Testing of Business-Cycle Theories, where Frisch accused Tinbergen of not having discovered autonomous relations, but "coflux" relations, [11] and by Jacob Marschak, in his 1953 contribution to the Cowles Commission Monograph, where he submitted that
The Lucas critique is representative of the paradigm shift that occurred in macroeconomic theory in the 1970s towards attempts at establishing micro-foundations.
In the 1980s, macro models emerged that attempted to directly respond to Lucas through the use of rational expectations econometrics. [12]
In 1982, Finn E. Kydland and Edward C. Prescott created a real business cycle (RBC) model to "predict the consequence of a particular policy rule upon the operating characteristics of the economy." [3] The stated, exogenous, stochastic components in their model are "shocks to technology" and "imperfect indicators of productivity." The shocks involve random fluctuations in the productivity level, which shift up or down the trend of economic growth. Examples of such shocks include innovations, the weather, sudden and significant price increases in imported energy sources, stricter environmental regulations, etc. The shocks directly change the effectiveness of capital and labour, which, in turn, affects the decisions of workers and firms, who then alter what they buy and produce. This eventually affects output. [3]
The authors stated that, since fluctuations in employment are central to the business cycle, the "stand-in consumer [of the model] values not only consumption but also leisure," meaning that unemployment movements essentially reflect the changes in the number of people who want to work. "Household-production theory," as well as "cross-sectional evidence" ostensibly support a "non-time-separable utility function that admits greater inter-temporal substitution of leisure, something which is needed," according to the authors, "to explain aggregate movements in employment in an equilibrium model." [3] For the K&P model, monetary policy is irrelevant for economic fluctuations.
The associated policy implications were clear: There is no need for any form of government intervention since, ostensibly, government policies aimed at stabilizing the business cycle are welfare-reducing. [12] Since microfoundations are based on the preferences of decision-makers in the model, DSGE models feature a natural benchmark for evaluating the welfare effects of policy changes. [13] [14] Furthermore, the integration of such microfoundations in DSGE modeling enables the model to accurately adjust to shifts in fundamental behaviour of agents and is thus regarded as an "impressive response" to the Lucas critique. [15] The Kydland/Prescott 1982 paper is often considered the starting point of RBC theory and of DSGE modeling in general [7] and its authors were awarded the 2004 Bank of Sweden Prize in Economic Sciences in Memory of Alfred Nobel. [16]
By applying dynamic principles, dynamic stochastic general equilibrium models contrast with the static models studied in applied general equilibrium models and some computable general equilibrium models.
DSGE models employed by governments and central banks for policy analysis are relatively simple. Their structure is built around three interrelated sections including that of demand, supply, and the monetary policy equation. These three sections are formally defined by micro-foundations and make explicit assumptions about the behavior of the main economic agents in the economy, i.e. households, firms, and the government. [17] The interaction of the agents in markets cover every period of the business cycle which ultimately qualifies the "general equilibrium" aspect of this model. The preferences (objectives) of the agents in the economy must be specified. For example, households might be assumed to maximize a utility function over consumption and labor effort. Firms might be assumed to maximize profits and to have a production function, specifying the amount of goods produced, depending on the amount of labor, capital and other inputs they employ. Technological constraints on firms' decisions might include costs of adjusting their capital stocks, their employment relations, or the prices of their products.
Below is an example of the set of assumptions a DSGE is built upon: [18]
to which the following frictions are added:
The models' general equilibrium nature is presumed to capture the interaction between policy actions and agents' behavior, while the models specify assumptions about the stochastic shocks that give rise to economic fluctuations. Hence, the models are presumed to "trace more clearly the shocks' transmission to the economy." [17] This is exemplified in the below explanation of a simplified DSGE model.
As such a complete simplified model of the relationship between three key features is defined. This dynamic interaction between the endogenous variables of output, inflation, and the nominal interest rate, is fundamental in DSGE modelling.
Two schools of analysis form the bulk of DSGE modeling: [note 4] the classic RBC models, and the New-Keynesian DSGE models that build on a structure similar to RBC models, but instead assume that prices are set by monopolistically competitive firms, and cannot be instantaneously and costlessly adjusted. Rotemberg & Woodford introduced this framework in 1997. Introductory and advanced textbook presentations of DSGE modeling are given by Galí (2008) and Woodford (2003). Monetary policy implications are surveyed by Clarida, Galí, and Gertler (1999).
The European Central Bank (ECB) has developed [19] a DSGE model, called the Smets–Wouters model, [20] which it uses to analyze the economy of the Eurozone as a whole. [note 5] The Bank's analysts state that
The main difference between "empirical" DSGE models and the "more traditional macroeconometric models, such as the Area-Wide Model", [21] according to the ECB, is that "both the parameters and the shocks to the structural equations are related to deeper structural parameters describing household preferences and technological and institutional constraints." [19] The Smets-Wouters model uses seven Eurozone area macroeconomic series: real GDP; consumption; investment; employment; real wages; inflation; and the nominal, short-term interest rate. Using Bayesian estimation and validation techniques, the bank's modeling is ostensibly able to compete with "more standard, unrestricted time series models, such as vector autoregression, in out-of-sample forecasting." [19]
Bank of Lithuania Deputy Chairman Raimondas Kuodis disputes the very title of DSGE analysis: The models, he claims, are neither dynamic (since they contain no evolution of stocks of financial assets and liabilities), stochastic (because we live in the world of Knightian uncertainty and, since future outcomes or possible choices are unknown, then risk analysis or expected utility theory are not very helpful), general (they lack a full accounting framework, a stock-flow consistent framework, which would significantly reduce the number of degrees of freedom in the economy), or even about equilibrium (since markets clear only in a few quarters). [22]
Willem Buiter, Citigroup Chief Economist, has argued that DSGE models rely excessively on an assumption of complete markets, and are unable to describe the highly nonlinear dynamics of economic fluctuations, making training in 'state-of-the-art' macroeconomic modeling "a privately and socially costly waste of time and resources". [23] Narayana Kocherlakota, President of the Federal Reserve Bank of Minneapolis, wrote that
N. Gregory Mankiw, regarded as one of the founders of New Keynesian DSGE modeling, has argued that
In the 2010 United States Congress hearings on macroeconomic modeling methods, held on 20 July 2010, and aiming to investigate why macroeconomists failed to foresee the 2007–2008 financial crisis, MIT professor of Economics Robert Solow criticized the DSGE models currently in use:
Commenting on the Congressional session, [26] The Economist asked whether agent-based models might better predict financial crises than DSGE models. [27]
Former Chief Economist and Senior Vice President of the World Bank Paul Romer [note 6] has criticized the "mathiness" of DSGE models [28] and dismisses the inclusion of "imaginary shocks" in DSGE models that ignore "actions that people take." [29] Romer submits a simplified [note 7] presentation of real business cycle (RBC) modelling, which, as he states, essentially involves two mathematical expressions: The well known formula of the quantity theory of money, and an identity that defines the growth accounting residual A as the difference between growth of output Y and growth of an index X of inputs in production.
Romer assigned to residual A the label "phlogiston" [note 8] while he criticized the lack of consideration given to monetary policy in DSGE analysis. [29] [note 9]
Joseph Stiglitz finds "staggering" shortcomings in the "fantasy world" the models create and argues that "the failure [of macroeconomics] were the wrong microfoundations, which failed to incorporate key aspects of economic behavior". He suggested the models have failed to incorporate "insights from information economics and behavioral economics" and are "ill-suited for predicting or responding to a financial crisis." [30] Oxford University's John Muellbauer put it this way: "It is as if the information economics revolution, for which George Akerlof, Michael Spence and Joe Stiglitz shared the Nobel Prize in 2001, had not occurred. The combination of assumptions, when coupled with the trivialisation of risk and uncertainty...render money, credit and asset prices largely irrelevant... [The models] typically ignore inconvenient truths." [31] Nobel laureate Paul Krugman asked, "Were there any interesting predictions from DSGE models that were validated by events? If there were, I'm not aware of it." [32]
Austrian economists reject DSGE modelling. Critique of DSGE-style macromodeling is at the core of Austrian theory, where, as opposed to RBC and New Keynesian models where capital is homogeneous [note 10] capital is heterogeneous and multi-specific and, therefore, production functions for the multi-specific capital are simply discovered over time. Lawrence H. White concludes [33] that present-day mainstream macroeconomics is dominated by Walrasian DSGE models, with restrictions added to generate Keynesian properties:
Hayek had criticized Wicksell for the confusion of thinking that establishing a rate of interest consistent with intertemporal equilibrium [note 11] also implies a constant price level. Hayek posited that intertemporal equilibrium requires not a natural rate but the "neutrality of money," in the sense that money does not "distort" (influence) relative prices. [34]
Post-Keynesians reject the notions of macro-modelling typified by DSGE. They consider such attempts as "a chimera of authority," [35] pointing to the 2003 statement by Lucas, the pioneer of modern DSGE modelling:
A basic Post Keynesian presumption, which Modern Monetary Theory proponents share, and which is central to Keynesian analysis, is that the future is unknowable and so, at best, we can make guesses about it that would be based broadly on habit, custom, gut-feeling, [note 12] etc. [35] In DSGE modeling, the central equation for consumption supposedly provides a way in which the consumer links decisions to consume now with decisions to consume later and thus achieves maximum utility in each period. Our marginal Utility from consumption today must equal our marginal utility from consumption in the future, with a weighting parameter that refers to the valuation that we place on the future relative to today. And since the consumer is supposed to always the equation for consumption, this means that all of us do it individually, if this approach is to reflect the DSGE microfoundational notions of consumption. However, post-Keynesians state that: no consumer is the same with another in terms of random shocks and uncertainty of income (since some consumers will spend every cent of any extra income they receive while others, typically higher-income earners, spend comparatively little of any extra income); no consumer is the same with another in terms of access to credit; not every consumer really considers what they will be doing at the end of their life in any coherent way, so there is no concept of a "permanent lifetime income", which is central to DSGE models; and, therefore, trying to "aggregate" all these differences into one, single "representative agent" is impossible. [35] These assumptions are similar to the assumptions made in the so-called Ricardian equivalence, whereby consumers are assumed to be forward looking and to internalize the government's budget constraints when making consumption decisions, and therefore taking decisions on the basis of practically perfect evaluations of available information. [35]
Extrinsic unpredictability, post-Keynesians state, has "dramatic consequences" for the standard, macroeconomic, forecasting, DSGE models used by governments and other institutions around the world. The mathematical basis of every DSGE model fails when distributions shift, since general-equilibrium theories rely heavily on ceteris paribus assumptions. [35] They point to the Bank of England's explicit admission [37] that none of the models they used and evaluated coped well during the 2007–2008 financial crisis, which, for the Bank, "underscores the role that large structural breaks can have in contributing to forecast failure, even if they turn out to be temporary."
Christian Mueller [38] points out that the fact that DSGE models evolve (see next section) constitutes a contradiction of the modelling approach in its own right and, ultimately, makes DSGE models subject to the Lucas critique. This contradiction arises because the economic agents in the DSGE models fail to account for the fact that the very models on the basis of which they form expectations evolve due to progress in economic research. While the evolution of DSGE models as such is predictable the direction of this evolution is not. In effect, Lucas' notion of the systematic instability of economic models carries over to DSGE models proving that they are not solving one of the key problems they are thought to be overcoming.
Federal Reserve Bank of Minneapolis president Narayana Kocherlakota acknowledges that DSGE models were "not very useful" for analyzing the 2007–2008 financial crisis but argues that the applicability of these models is "improving," and claims that there is growing consensus among macroeconomists that DSGE models need to incorporate both "price stickiness and financial market frictions." [6] Despite his criticism of DSGE modelling, he states that modern models are useful:
Still, Kocherlakota observes that in "terms of fiscal policy (especially short-term fiscal policy), modern macro-modeling seems to have had little impact. ... [M]ost, if not all, of the motivation for the fiscal stimulus was based largely on the long-discarded models of the 1960s and 1970s. [6]
In 2010, Rochelle M. Edge, of the Federal Reserve System Board of Directors, contested that the work of Smets & Wouters has "led DSGE models to be taken more seriously by central bankers around the world" so that "DSGE models are now quite prominent tools for macroeconomic analysis at many policy institutions, with forecasting being one of the key areas where these models are used, in conjunction with other forecasting methods." [39]
University of Minnesota professor of economics V.V. Chari has pointed out that state-of-the-art DSGE models are more sophisticated than their critics suppose:
Chari also argued that current DSGE models frequently incorporate frictional unemployment, financial market imperfections, and sticky prices and wages, and therefore imply that the macroeconomy behaves in a suboptimal way which monetary and fiscal policy may be able to improve. [40] Columbia University's Michael Woodford concedes [41] that policies considered by DSGE models might not be Pareto optimal [note 14] and they may as well not satisfy some other social welfare criterion. Nonetheless, in replying to Mankiw, Woodford argues that the DSGE models commonly used by central banks today and strongly influencing policy makers like Ben Bernanke, do not provide an analysis so different from traditional Keynesian analysis:
Macroeconomics is a branch of economics that deals with the performance, structure, behavior, and decision-making of an economy as a whole. This includes regional, national, and global economies. Macroeconomists study topics such as output/GDP and national income, unemployment, price indices and inflation, consumption, saving, investment, energy, international trade, and international finance.
Rational expectations is an economic theory that seeks to infer the macroeconomic consequences of individuals' decisions based on all available knowledge. It assumes that individuals' actions are based on the best available economic theory and information.
New Keynesian economics is a school of macroeconomics that strives to provide microeconomic foundations for Keynesian economics. It developed partly as a response to criticisms of Keynesian macroeconomics by adherents of new classical macroeconomics.
This aims to be a complete article list of economics topics:
Business cycles are intervals of general expansion followed by recession in economic performance. The changes in economic activity that characterize business cycles have important implications for the welfare of the general population, government institutions, and private sector firms.
The Lucas critique argues that it is naïve to try to predict the effects of a change in economic policy entirely on the basis of relationships observed in historical data, especially highly aggregated historical data. More formally, it states that the decision rules of Keynesian models—such as the consumption function—cannot be considered as structural in the sense of being invariant with respect to changes in government policy variables. It was named after American economist Robert Lucas's work on macroeconomic policymaking.
A macroeconomic model is an analytical tool designed to describe the operation of the problems of economy of a country or a region. These models are usually designed to examine the comparative statics and dynamics of aggregate quantities such as the total amount of goods and services produced, total income earned, the level of employment of productive resources, and the level of prices.
Neutrality of money is the idea that a change in the stock of money affects only nominal variables in the economy such as prices, wages, and exchange rates, with no effect on real variables, like employment, real GDP, and real consumption. Neutrality of money is an important idea in classical economics and is related to the classical dichotomy. It implies that the central bank does not affect the real economy by creating money. Instead, any increase in the supply of money would be offset by a proportional rise in prices and wages. This assumption underlies some mainstream macroeconomic models. Others like monetarism view money as being neutral only in the long run.
John Brian Taylor is the Mary and Robert Raymond Professor of Economics at Stanford University, and the George P. Shultz Senior Fellow in Economics at Stanford University's Hoover Institution.
Microfoundations are an effort to understand macroeconomic phenomena in terms of economic agents' behaviors and their interactions. Research in microfoundations explores the link between macroeconomic and microeconomic principles in order to explore the aggregate relationships in macroeconomic models.
The neoclassical synthesis (NCS), or neoclassical–Keynesian synthesis is an academic movement and paradigm in economics that worked towards reconciling the macroeconomic thought of John Maynard Keynes in his book The General Theory of Employment, Interest and Money (1936) with neoclassical economics.
New classical macroeconomics, sometimes simply called new classical economics, is a school of thought in macroeconomics that builds its analysis entirely on a neoclassical framework. Specifically, it emphasizes the importance of rigorous foundations based on microeconomics, especially rational expectations.
The AD–AS or aggregate demand–aggregate supply model is a widely used macroeconomic model that explains short-run and long-run economic changes through the relationship of aggregate demand (AD) and aggregate supply (AS) in a diagram. It coexists in an older and static version depicting the two variables output and price level, and in a newer dynamic version showing output and inflation.
Jordi Galí is a Catalan macroeconomist who is regarded as one of the main figures in New Keynesian macroeconomics today. He is a Senior Researcher at the Centre de Recerca en Economia Internacional (CREI), a Professor at Universitat Pompeu Fabra and a Research Professor at the Barcelona School of Economics. After obtaining his doctorate from MIT in 1989 under the supervision of Olivier Blanchard, he held faculty positions at Columbia University and New York University before moving to Barcelona.
Macroeconomic theory has its origins in the study of business cycles and monetary theory. In general, early theorists believed monetary factors could not affect real factors such as real output. John Maynard Keynes attacked some of these "classical" theories and produced a general theory that described the whole economy in terms of aggregates rather than individual, microeconomic parts. Attempting to explain unemployment and recessions, he noticed the tendency for people and businesses to hoard cash and avoid investment during a recession. He argued that this invalidated the assumptions of classical economists who thought that markets always clear, leaving no surplus of goods and no willing labor left idle.
Following the development of Keynesian economics, applied economics began developing forecasting models based on economic data including national income and product accounting data. In contrast with typical textbook models, these large-scale macroeconometric models used large amounts of data and based forecasts on past correlations instead of theoretical relations. These models estimated the relations between different macroeconomic variables using regression analysis on time series data. These models grew to include hundreds or thousands of equations describing the evolution of hundreds or thousands of prices and quantities over time, making computers essential for their solution. While the choice of which variables to include in each equation was partly guided by economic theory, variable inclusion was mostly determined on purely empirical grounds. Large-scale macroeconometric model consists of systems of dynamic equations of the economy with the estimation of parameters using time-series data on a quarterly to yearly basis.
The new neoclassical synthesis (NNS), which is occasionally referred as the New Consensus, is the fusion of the major, modern macroeconomic schools of thought – new classical macroeconomics/real business cycle theory and early New Keynesian economics – into a consensus view on the best way to explain short-run fluctuations in the economy. This new synthesis is analogous to the neoclassical synthesis that combined neoclassical economics with Keynesian macroeconomics. The new synthesis provides the theoretical foundation for much of contemporary mainstream macroeconomics. It is an important part of the theoretical foundation for the work done by the Federal Reserve and many other central banks.
Real business-cycle theory is a class of new classical macroeconomics models in which business-cycle fluctuations are accounted for by real shocks. Unlike other leading theories of the business cycle, RBC theory sees business cycle fluctuations as the efficient response to exogenous changes in the real economic environment. That is, the level of national output necessarily maximizes expected utility, and governments should therefore concentrate on long-run structural policy changes and not intervene through discretionary fiscal or monetary policy designed to actively smooth out economic short-term fluctuations.
A Calvo contract is the name given in macroeconomics to the pricing model that when a firm sets a nominal price there is a constant probability that a firm might be able to reset its price which is independent of the time since the price was last reset. The model was first put forward by Guillermo Calvo in his 1983 article "Staggered Prices in a Utility-Maximizing Framework". The original article was written in a continuous time mathematical framework, but nowadays is mostly used in its discrete time version. The Calvo model is the most common way to model nominal rigidity in new Keynesian DSGE macroeconomic models.
Stephanie Schmitt-Grohé is a German economist who has been a professor of economics at Columbia University since 2008. Her research focuses on macroeconomics, fiscal policy, and monetary policy in open and closed economies. In 2004, she was awarded the Bernácer Prize, for her research on monetary stabilization policies.