What big data can tell us about the cost of living

A Businesswoman is silhouetted as she makes her way under the Arche de la Defense, in the financial district west of Paris, November 20, 2012. France said its economy was sound and reforms were on track after credit ratings agency Moody's stripped it of the prized triple-A badge due to an uncertain fiscal and economic outlook. Monday's downgrade, which follows a cut by Standard & Poor's in January, was expected but is a blow to Socialist President Francois Hollande as he tries to fix France's finances and revive the euro zone's second largest economy.

The precision of the data impels us to be precise about the theoretical assumptions. Image: REUTERS/Christian Hartmann

David E. Weinstein
Professor, Center for Japanese Economy and Business
Stephen Redding
Harold T. Shapiro Professor in Economics, Economics Department and Woodrow Wilson School, Princeton University

The usage of big data is likely to transform economic measurement in ways that we are only beginning to grasp (Cavallo and Rigobon 2016). Big data encompasses four fundamental shifts from standard datasets: volume, veracity, velocity, and variety. It is commonplace to focus on the first three of these. Clearly, scanner data offers us more observations, fewer opportunities for human input errors to creep in, and the capacity to measure prices and sales almost instantaneously instead of waiting for monthly surveys. These benefits are creating opportunities for improving the timeliness and quality of existing measurement techniques. However, in a new paper, we argue that the most exciting part of the big data revolution is likely to come from the new varieties of data that have become available (Redding and Weinstein 2016).

One of the principal challenges in producing numbers like real GDP or real wages is that while nominal variables are easy to measure, the measurement of real variables requires a theory of economic behaviour. Just as accountants like to joke that “sales are a fact; profits are an idea”, in economics, we face a similar conundrum -“consumer expenditures are a fact; real income is an idea”. While few people would disagree about what the nominal sales of any firm are or how much a consumer spends on a product, translating nominal numbers into real output or welfare is challenging (and was a key component of the path-breaking work of the recent Nobel Laureate, Angus Deaton, as in Deaton and Muellbauer 1980).

Unfortunately, the problem is not that we don’t know how to convert nominal expenditures into welfare; it is that we know too many ways of doing it. Broadly speaking, the profession has settled on three disjoint approaches. First, macroeconomists typically assume there are no demand shifts when measuring real income movements. A foundational assumption in these models is the idea that taste parameters never shift, so the utility function is constant. Economists make this assumption in order to derive a ‘money-metric’ utility function, which guarantees that welfare can be measured if one only knows income and prices. Applied microeconomists take a very different approach by assuming that there are time-varying demand and supply curves. Although it is not typically acknowledged, the existence of these time-varying demand curves is inconsistent with the macroeconomist’s idea that taste parameters are fixed. Demand shifts reflect the fact that a consumer likes one product more than another, which in general will mean that utility is not money-metric. Finally, actual price and real output data is constructed by statistical agencies using formulas that differ from either approach.

The inconsistencies are so deep that the same assumptions that form the foundation of demand-system estimation can be used to prove that standard price indexes are incorrect, and the assumptions underlying standard price indexes invalidate demand-system estimation because if no demand parameter ever shifts, one can recover the demand elasticity without recourse to estimation. In other words, extant micro and macro welfare estimates are inconsistent with each other as well as the data.

One can ignore these problems in conventional datasets, because one typically does not observe all the key features of each transaction. Thus, one can assume errors in demand systems are due to unobserved quality changes, measurement error, or something other than a shift in demand. However, it is much harder to ignore these contradictions in barcode data because we observe prices and quantities sold of precisely defined products. While conventional price measurement is based on identifying ‘the price’ for heterogeneous bundles of goods (milk, carbonated beverages, computers) that contain substantial quality variation within product categories, barcode data eliminates the ambiguity in the definition of a product. Put differently, while one might be able to assume that shifts in demand for carbonated beverages are the result of unobserved quality upgrading, it strains credulity to make the same assumption for a 330mL can of Coke Zero.

The precision of the data impels us to be precise about the theoretical assumptions. In order to deal with this problem, we present a new empirical methodology, which we term ‘the unified approach’, that nests all major price indexes used in welfare or demand system analysis (Redding and Weinstein 2016). We show that the measures of welfare used by economists and statistical agencies can be understood in terms of an internally consistent approach that has been altered by ignoring data or theoretical conditions that must hold in any coherent system. As shown in Figure 1, all of the major approaches to price (and therefore welfare) measurement are actually linked together via the unified approach.

Figure 1. The big picture

 The big picture
Image: VoxEU

The first key insight of our unified approach is that any demand system errors (e.g. taste shocks) must show up in the utility and unit expenditure functions, and therefore the price index. However, all economically motivated price indexes are derived under the assumption that the demand parameter for each good is time invariant. Researchers make this assumption because it is a sufficient condition to guarantee the existence of a constant aggregate utility function, but it stands in the face of an overwhelming body of evidence that demand curves shift.

Our analysis shows that the assumption of time-invariant preferences for each good is neither the correct nor necessary condition to make consistent comparisons of welfare over time when there are demand shocks for individual goods. To be able to make such consistent welfare comparisons, one must obtain the same change in the cost of living between a pair of time periods, whether one uses today's preferences for both periods, yesterday's preferences for both periods, or the preferences for each period. This necessary condition is (trivially) satisfied when preferences for each good are time invariant. One of our main contributions is to show that this necessary conditional can also be satisfied when shifts in demand cancel on average. This yields our ‘unified price index’ that is valid even when the set of goods is changing over time (due to product innovation, as in Feenstra 1994).

Our index enables us to identify a novel form of bias that arises from the assumption of time-invariant demand for each good in existing price indexes. ‘Consumer valuation bias’ arises whenever expenditure shares respond to demand shifts. Since conventional indexes assume that expenditure shares are only affected by price changes, they will be biased whenever expenditure share changes are correlated with demand shifts. For example, if higher consumer demand causes prices to rise, a conventional index will overstate cost-of-living changes because it will not adjust for the fact that some of the price increase is offset by the higher utility per unit associated with the demand shift.

Our second main insight is to develop a novel way of estimating the elasticity of substitution between goods. Extant approaches focus on identification from supply and demand systems. However, we show that one can also identify this parameter by combining information from the demand system and unit expenditure function. One of the desirable properties of this ‘reverse-weighting’ estimator is that it minimises departures from money-metric utility given the observed data on prices and expenditure and the assumed constant elasticity of substitution utility function.

Finally, we use barcode data to examine the properties of our unified price index and reverse-weighting estimator. We find that we obtain reasonable elasticity estimates in the sense that they are similar to those identified using other methodologies on the same data. Moreover, the consumer valuation biases in existing indexes appear to be quite substantial, suggesting that allowing for demand shifts is an economically important force in understanding price and real income changes.

Figure 2. Changes in cost of living for various indexes

 Changes in cost of living for various indexes
Image: VoxEU

We can see these differences at the aggregate level in Figure 2, which plots the expenditure-share-weighted average of the changes in the cost of living across product groups for each of the different index numbers over time, again using the initial period expenditure share weights. Not surprisingly, the Fisher, Törnqvist and Sato-Vartia result in almost identical changes in the cost of living that are bounded by the Paasche and Laspeyres indexes. This similarity is driven by the fact that they all assume no demand shifts for any good. The distance between the Sato-Vartia index and the Common-Goods Unified Price Index tells us the importance of the consumer valuation bias and the distance between the Sato-Vartia and the Feenstra-CPI indicates the value of the adjustment for changes in variety. In other words, big data suggests that standard methods of measuring welfare overstate cost of living increases by several percentage points per year because they ignore new goods and demand shifts.

References

Cavallo, A and R Rigobon (2016) “The Billion Prices Project: Using online data for measurement and research”, VoxEU, 24 April.

Deaton, A and J Muellbauer (1980) Economics and Consumer Behavior, Cambridge: Cambridge University Press.

Feenstra, R C (1994) “New product varieties and the measurement of international prices,”American Economic Review, 84(1): 157-177.

Redding, S J and D E Weinstein (2016) “A unified approach to estimating demand and welfare,” NBER Working Paper, no 22479.

Don't miss any update on this topic

Create a free account and access your personalized content collection with our latest publications and analyses.

Sign up for free

License and Republishing

World Economic Forum articles may be republished in accordance with the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License, and in accordance with our Terms of Use.

The views expressed in this article are those of the author alone and not the World Economic Forum.

Stay up to date:

Data Science

Share:
The Big Picture
Explore and monitor how Data Science is affecting economies, industries and global issues
World Economic Forum logo

Forum Stories newsletter

Bringing you weekly curated insights and analysis on the global issues that matter.

Subscribe today

About us

Engage with us

  • Sign in
  • Partner with us
  • Become a member
  • Sign up for our press releases
  • Subscribe to our newsletters
  • Contact us

Quick links

Language editions

Privacy Policy & Terms of Service

Sitemap

© 2024 World Economic Forum