In this paper we investigate the role of input-output data source in the regional econometric input-output models. While there has been a great deal of experimentation focused on the accuracy of alternative methods for estimating regional input-output coefficients, little attention has been directed to the role of accuracy when the input-output system is nested within a broader accounting framework. The issues of accuracy were considered in two contexts, forecasting and impact analysis focusing on a model developed for the Chicago Region. We experimented with three input-output data sources: observed regional data, national input-output, and randomly generated input-output coefficients. The effects of different sources of input-output data on regional econometric input-output model revealed that there are significant differences in results obtained in impact analyses. However, the adjustment processes inherent in the econometric input-output system seem to mute the initial differences in input- output data when the model is used for forecasting. Since applications of these types of models involve both impact and forecasting exercises, there would still seem to be a strong motivation for basing the system on the most accurate set of input-output accounts.
-
Measuring Plant Level Energy Efficiency and Technical Change in the U.S. Metal-Based Durable Manufacturing Sector Using Stochastic Frontier Analysis
January 2016
Working Paper Number:
CES-16-52
This study analyzes the electric and thermal energy efficiency for five different metal-based durable manufacturing industries in the United States from 1987-2012 at the 3 digit North American Industry Classification System (NAICS) level. Using confidential plant-level data on energy use and production from the quinquennial U.S. Economic Census, a stochastic frontier regression analysis (SFA) is applied in six repeated cross sections for each five year census. The SFA controls for energy prices and climate-driven energy demand (heating degree days - HDD - and cooling degree days - CDD) due to differences in plant level locations, as well as 6-digit NAICS industry effects. A Malmquist index is used to decompose aggregate plant technical change in energy use into indices of efficiency and frontier (best practice) change. Own energy price elasticities range from -.7 to -1.0, with electricity tending to have slightly higher elasticity than fuel. Mean efficiency estimates (100 percent equals best practice level) range from a low of 32 percent (thermal 334 - Computer and Electronic Products) to a high of 86 percent (electricity 332 - Fabricated Metal Products). Electric efficiency is consistently better than thermal efficiency for all NAICS. There is no clear pattern to the decomposition of aggregate technical Thermal change. In some years efficiency improvement dominates; in other years aggregate technical change is driven by improvement in best practice.
View Full
Paper PDF
-
CONSTRUCTION OF REGIONAL INPUT-OUTPUT TABLES FROM ESTABLISHMENT-LEVEL MICRODATA: ILLINOIS, 1982
August 1993
Working Paper Number:
CES-93-12
This paper presents a new method for use in the construction of hybrid regional input-output tables, based primarily on individual returns from the Census of Manufactures. Using this method, input- output tables can be completed at a fraction of the cost and time involved in the completion of a full survey table. Special attention is paid to secondary production, a problem often ignored by input-output analysts. A new method to handle secondary production is presented. The method reallocates the amount of secondary production and its associated inputs, on an establishment basis, based on the assumption that the input structure for any given commodity is determined not by the industry in which the commodity was produced, but by the commodity itself -- the commodity-based technology assumption. A biproportional adjustment technique is used to perform the reallocations.
View Full
Paper PDF
-
The Spatial Extent of Agglomeration Economies: Evidence from Three U.S. Manufacturing Industries
January 2012
Working Paper Number:
CES-12-01
The spatial extent of localized agglomeration economies constitutes one of the central current questions in regional science. It is crucial for understanding firm location decisions and for assessing the influence of proximity in shaping spatial patterns of economic activity, yet clear-cut answers are difficult to come by. Theoretical work often fails to define or specify the spatial dimension of agglomeration phenomena. Existing empirical evidence is far from consistent. Most sources of data on economic performance do not supply micro-level information containing usable geographic locations. This paper provides evidence of the distances across which distinct sources of agglomeration economies generate benefits for plants belonging to three manufacturing industries in the United States. Confidential data from the Longitudinal Research Database of the United States Census Bureau are used to estimate cross-sectional production function systems at the establishment level for three contrasting industries in three different years. Along with relevant establishment, industry, and regional characteristics, the production functions include variables that indicate the local availability of potential labor and supply pools and knowledge spillovers. Information on individual plant locations at the county scale permits spatial differentiation of the agglomeration variables within geographic regions. Multiple distance decay profiles are investigated in order to explore how modifying the operationalization of proximity affects indicated patterns of agglomeration externalities and interfirm interactions. The results imply that industry characteristics are at least as important as the type of externality mechanism in determining the spatial pattern of agglomeration benefits. The research methods borrow from earlier work by the author that examines the relationships between regional industrial structure and manufacturing production.
View Full
Paper PDF
-
Building the Census Bureau Index of Economic Activity (IDEA)
March 2023
Working Paper Number:
CES-23-15
The Census Bureau Index of Economic Activity (IDEA) is constructed from 15 of the Census Bureau's primary monthly economic time series. The index is intended to provide a single time series reflecting, to the extent possible, the variation over time in the whole set of component series. The component series provide monthly measures of activity in retail and wholesale trade, manufacturing, construction, international trade, and business formations. Most of the input series are Principal Federal Economic Indicators. The index is constructed by applying the method of principal components analysis (PCA) to the time series of monthly growth rates of the seasonally adjusted component series, after standardizing the growth rates to series with mean zero and variance 1. Similar PCA approaches have been used for the construction of other economic indices, including the Chicago Fed National Activity Index issued by the Federal Reserve Bank of Chicago, and the Weekly Economic Index issued by the Federal Reserve Bank of New York. While the IDEA is constructed from time series of monthly data, it is calculated and published every business day, and so is updated whenever a new monthly value is released for any of its component series. Since release dates of data values for a given month vary across the component series, with slight variations in the monthly release date for any one component series, updates to the index are frequent. It is unavoidably the case that, at almost all updates, some of the component series lack observations for the current (most recent) data month. To address this situation, component series that are one month behind are predicted (nowcast) for the current index month, using a multivariate autoregressive time series model. This report discusses the input series to the index, the construction of the index by PCA, and the nowcasting procedure used. The report then examines some properties of the index and its relation to quarterly U.S. Gross Domestic Product and to some monthly non-Census Bureau economic indicators.
View Full
Paper PDF
-
Who are the people in my neighborhood? The 'contextual fallacy' of measuring individual context with census geographies
February 2018
Working Paper Number:
CES-18-11
Scholars deploy census-based measures of neighborhood context throughout the social sciences and epidemiology. Decades of research confirm that variation in how individuals are aggregated into geographic units to create variables that control for social, economic or political contexts can dramatically alter analyses. While most researchers are aware of the problem, they have lacked the tools to determine its magnitude in the literature and in their own projects. By using confidential access to the complete 2010 U.S. Decennial Census, we are able to construct'for all persons in the US'individual-specific contexts, which we group according to the Census-assigned block, block group, and tract. We compare these individual-specific measures to the published statistics at each scale, and we then determine the magnitude of variation in context for an individual with respect to the published measures using a simple statistic, the standard deviation of individual context (SDIC). For three key measures (percent Black, percent Hispanic, and Entropy'a measure of ethno-racial diversity), we find that block-level Census statistics frequently do not capture the actual context of individuals within them. More problematic, we uncover systematic spatial patterns in the contextual variables at all three scales. Finally, we show that within-unit variation is greater in some parts of the country than in others. We publish county-level estimates of the SDIC statistics that enable scholars to assess whether mis-specification in context variables is likely to alter analytic findings when measured at any of the three common Census units.
View Full
Paper PDF
-
Estimating the Distribution of Plant-Level Manufacturing Energy Efficiency with Stochastic Frontier Regression
March 2007
Working Paper Number:
CES-07-07
A feature commonly used to distinguish between parametric/statistical models and engineering models is that engineering models explicitly represent best practice technologies while the parametric/statistical models are typically based on average practice. Measures of energy intensity based on average practice are less useful in the corporate management of energy or for public policy goal setting. In the context of company or plant level energy management, it is more useful to have a measure of energy intensity capable of representing where a company or plant lies within a distribution of performance. In other words, is the performance close (or far) from the industry best practice? This paper presents a parametric/statistical approach that can be used to measure best practice, thereby providing a measure of the difference, or 'efficiency gap' at a plant, company or overall industry level. The approach requires plant level data and applies a stochastic frontier regression analysis to energy use. Stochastic frontier regression analysis separates the energy intensity into three components, systematic effects, inefficiency, and statistical (random) error. The stochastic frontier can be viewed as a sub-vector input distance function. One advantage of this approach is that physical product mix can be included in the distance function, avoiding the problem of aggregating output to define a single energy/output ratio to measure energy intensity. The paper outlines the methods and gives an example of the analysis conducted for a non-public micro-dataset of wet corn refining plants.
View Full
Paper PDF
-
Environmental Regulation, Abatement, and Productivity: A Frontier Analysis
September 2013
Working Paper Number:
CES-13-51
This research studies the link between environmental regulation and plant level productivity in two U.S. manufacturing industries: pulp and paper mills and oil refineries using Data Envelopment Analysis (DEA) models. Data on abatement spending, emissions and abated emissions are used in different DEA models to study plant productivity outcomes when accounting for abatement spending or emissions regulations. Results indicate that pulp and paper mills and oil refineries in the U.S. suffered decreases in productivity due to pollution abatement activities from 1974 to 2000. These losses in productivity are substantial but have been slowly trending downwards even when the regulations have tended to become more stringent and emission of pollutants has declined suggesting that the best practice has shifted over time. Results also show that the reported abatement expenditures are not able to explain all the losses arising out of regulation suggesting that these abatement expenditures are consistently under-reported.
View Full
Paper PDF
-
An Applied General Equilibrium Model Of Moroccan Trade Liberalization Featuring External Economies
November 1997
Working Paper Number:
CES-97-16
Since the 1920's economists have wrestled with the effects of external economies on trade liberalization. In this paper I show that under extreme conditions, externalities can reverse the gains from trade found in perfectly competitive trade models. However, the externalities needed to generate this result, even under the worst possible conditions (all expanding industries are subject to negative externalities, all contracting industries have positive externalities) are orders of magnitude larger than those estimated in Krizan (1997). This suggests that the presence of external economies of scale does not provide a credible argument for protectionism. On the other hand, the CGE model showed that external effects can increase the welfare gains from trade liberalization, but the combined effect is still small compared to other policy options. This finding contrasts sharply with many models featuring internal returns to scale that are able to generate large welfare benefits from trade liberalization.
View Full
Paper PDF
-
Decomposing Learning By Doing in New Plants
December 1992
Working Paper Number:
CES-92-16
The paper examines learning by doing in the context of a production function in which the other arguments are labor, human capital, physical capital, and vintage as a proxy for embodied technical change in physical capital. Learning is further decomposed into organization learning, capital learning, and manual task learning. The model is tested with time series and cross section data for various samples of up to 2,150 plants over a 14 year period. Word Perfect Version
View Full
Paper PDF
-
Capital-Energy Substitution Revisted: New Evidence From Micro Data
April 1997
Working Paper Number:
CES-97-04
We use new micro data for 11,520 plants taken from the Census Bureau=s 1991 Manufacturing Energy Consumption Survey (MECS) and 1991 Annual Survey of Manufactures (ASM) to estimate elasticities of substitution between energy and capital. We found that energy and capital are substitutes. We also found that estimates of Allen elasticities of substitution -- which have been used as a standard measure of substitution -- are sensitive to varying data sets and levels of aggregation. In contrast, estimates of Morishima elasticities of substitution -- which are theoretically superior to the Allen elasticities -- are more robust (except when two-digit level data are used). The results support the views that (i) the Morishima elasticity is a better measure of factor substitution and (ii) micro data provide more accurate elasticity estimates than those obtained from aggregate data. Our findings appear to resolve the long-standing conflict among the estimates reported in the many previous studies regarding energy-capital substitution/complementarity.
View Full
Paper PDF