Measuring U.S. Innovative Activities

38
1 Measuring U.S. Innovative Activities 2008 World Congress on NAEP Measures for Nations Washington, DC May 16, 2008 B.K. Atrostic, Senior Economist Center for Economic Studies, U.S. Census Bureau [email protected] v

description

Measuring U.S. Innovative Activities. 2008 World Congress on NAEP Measures for Nations Washington, DC May 16, 2008 B.K. Atrostic, Senior Economist Center for Economic Studies, U.S. Census Bureau. [email protected]. Disclaimer. - PowerPoint PPT Presentation

Transcript of Measuring U.S. Innovative Activities

Page 1: Measuring U.S. Innovative Activities

1

Measuring U.S. Innovative Activities

2008 World Congress on NAEP Measures for Nations

Washington, DC

May 16, 2008

B.K. Atrostic, Senior EconomistCenter for Economic Studies, U.S. Census Bureau

[email protected]

Page 2: Measuring U.S. Innovative Activities

2

Disclaimer

The views expressed in this presentation are those of the author and not necessarily those of the U.S. Census Bureau.

Page 3: Measuring U.S. Innovative Activities

3

Overview

• Importance of Measuring Innovative Activity

• Allocating Scarce Statistical Resources

• Definitions

• U.S.: Data & Lessons Learned

• Other Countries: Data & Lessons Learned

• Critical U.S. Data Gaps – Identifying and Filling

• Innovative Activities and Intangibles

Page 4: Measuring U.S. Innovative Activities

4

Why Measure Innovative Activity?

• Identify sources of new goods and services

• Identify sources of total factor productivity, growth, other economic outcomes

Page 5: Measuring U.S. Innovative Activities

5

Measuring Innovative Activity and Its Impact

• Innovation surveys – European countries, Canada, Australia, many

other countries– Georgia, for the U.S.

• Fill key gaps in core microdata needed to analyze economic performance

Page 6: Measuring U.S. Innovative Activities

6

Scarce Statistical System Resources

• Focus on U.S. Census Bureau Data Collections– Dollars– Expertise / core competencies– Respondent burden / cooperation

Page 7: Measuring U.S. Innovative Activities

7

Filling Gaps in Core Data

• Longitudinal data are critical Without longitudinal data, it is hard to rule out the

possibility that a measure of innovative activity merely proxies for something omitted from or measured poorly in the core data.

• Similar issues in developing new measures of effects of information technology on economic performance

• Neither innovation survey nor more systematic data on innovative activity would fill core data gaps

Page 8: Measuring U.S. Innovative Activities

8

Innovation: Definitions• OECD “Oslo” Innovation Manual

– Guidelines for Collecting and Interpreting Innovation Data

– 1992 / 1997 / 2005 editions

– 163 pages in 2005 edition

• Advisory Committee on Measuring Innovation in the 21st Century Economy– Established by U.S. Secretary of Commerce in 2006

• Related but different idea: Research and Development (R&D)– OECD “Frascati” R&D manual 1963 / 1993 / 2002

Page 9: Measuring U.S. Innovative Activities

9

Innovation Definitions • Oslo Manual:

– The implementation of a new or significantly improved product (good or service), or process, a new marketing method, or a new organizational method in business practices, workplace organization, or external relations (2005 version, p. 46).

• Advisory Committee on Measuring Innovation in the 21st Century Economy– The design, invention, development and/or implementation of new or

altered products, services, processes, systems, organizational structures, or business models for the purpose of creating new value for customers and financial returns for the firm (January 2008, p. I).

Page 10: Measuring U.S. Innovative Activities

10

Innovative Activity Defined

• Related to but distinct from R&D and innovation

• Specific identifiable activity– Uses products of innovation process from own or other

industries or sectors

– Provides environment conducive to innovation,

– Serves as marker of innovative managerial or organizational capacity

Page 11: Measuring U.S. Innovative Activities

11

Innovative Activity: Data and Research Findings

U.S. Census Bureau Data• Research Conducted at Center for Economic

Studies or Census Bureau Research Data Centers– Diffusion of Innovation– Human and Organizational Capital– Other Topics

• Findings summarized in Atrostic 2007 and 2008

Page 12: Measuring U.S. Innovative Activities

12

Innovative Activity: Data and Research Findings

Other Countries• Community Innovation Survey (CIS)

• Research Findings

Page 13: Measuring U.S. Innovative Activities

13

U.S. Census Bureau Data: Diffusion of Innovation

• Survey of Manufacturing Technologies: 1988

and 1993 ASM supplements

• Computer Network Use Survey (CNUS): 1999

ASM supplement

• Information and Communications Technology Survey (ICTS): 2003 + Annual Capital Expenditure

Survey (ACES) Supplements

Page 14: Measuring U.S. Innovative Activities

14

U.S. Census Bureau Data: Human and Organizational

Capital• Broader capital measures would include richer

measures of a firm’s human and organizational capital (Corrado et al. 2005, 2006)

• National Employer Surveys (NES): 1994, 1996, 1997, 1998,

2000

• Entrepreneurship: – Integrated Longitudinal Business Database (ILBD),

– Characteristics of Business Owners Survey (CBO)

• Firm Organization

Page 15: Measuring U.S. Innovative Activities

15

U.S. Census Bureau Data: Other Topics

• Dynamics of Labor Markets:– Worker and Firm Characteristics: Longitudinal Employer-

Household Dynamics (LEHD)

• International Trade:– Importing, Exporting, Outsourcing

– Foreign Direct Investment • Successful BEA/NSF/Census Bureau pilot project

• Matched BEA FDI and U.S. Direct Investment Abroad with NSF/Census Survey of Industrial Research and Development (SIRD)

• Subsequent extension and expansion of pilot

Page 16: Measuring U.S. Innovative Activities

16

Innovation Measurement in Other Countries

• Community Innovation Survey (CIS) – Eurostat sponsored (required of EU countries)– 1992, 1996, 2001, 2003, 2005, 2008– Based on Oslo manual– CIS 3 & 4 harmonized core questions– European Innovation Scorecard publication– Similar surveys in many non-EU countries– Lively microdata research

Page 17: Measuring U.S. Innovative Activities

17

Innovation Measurement: Canada

• Since 1993

• Also survey advanced technologies and practices, not consistent over time

• Based on Oslo manual

• Decade of research links innovation measures with economic inputs, outcomes

• http://www.statcan.ca/english/freepub/11-623-XIE/2003001/innov.htm

• Innovation Analysis Bulletin publication

Page 18: Measuring U.S. Innovative Activities

18

Innovation Measurement in Other Countries

• Australia:– Innovation surveys 1992-1993, 1996-1997, 2003-2004, 2005

• 2003-2004 not based on CIS

– Integrated Business Characteristics Strategy• Core IT and innovation questions annually

• Detailed IT or innovation questions alternate years

• Links to longitudinal business database

• IT use and Innovation, 2005-2006, released 2007

• Selected characteristics of businesses, 2005-2006, released 2008.

Page 19: Measuring U.S. Innovative Activities

19

Innovation Measurement in Georgia

• Georgia Manufacturing Survey– Georgia Tech’s Enterprise Innovation Institute &

– Georgia Tech School of Public Policy

• 1994, 1996, 1999, 2002, 2005, 2008

• Many questions designed and worded to parallel CIS

• Concurrent Workshop May 12 – 13, 2008– International microdata – Selected CIS countries

– Parallel studies, collaborative research

Page 20: Measuring U.S. Innovative Activities

20

International Lessons Learned Analytical Findings

• Hall and Mairesse 2006 Economics of Innovation and New

Technology

• Jaumotte and Pain 2005

• Parvan 2007

• OECD NESTI / WPIA innovation microdata project– Internationally comparable econometric analyses

– Workshops 2007 – 2008

• Related internationally comparable microdata work– UK-led ICT impact analyses

Page 21: Measuring U.S. Innovative Activities

21

International Lessons Learned Not Just Innovation

– Canada: Dourfour, Nakamura, Tang 2006• Advanced technology and business practices

• Significant interactions

– UK: Bloom, Sadum, Van Reenen 2006• IT, international trade, ownership (UK vs. not)

• Significant interactions

– Germany & Netherlands: Hempell, van Leeuwen, van der Wiel 2006

• Innovation significant in MFP

• Significant interactions with IT

Page 22: Measuring U.S. Innovative Activities

22

International Lessons Learned “Best Practice”

• Current “State of Art” Not Settled:– Australia’s indicators and structures– EU-like CIS: Empirical and econometric

properties of innovation data being tested

• Can CIS be applied straightforwardly to U.S.?

Page 23: Measuring U.S. Innovative Activities

23

Methodological Issues in Implementing Best Practices

• Committee on National Statistics 2005– Innovation, linked activities, and outcomes can be measured

– Many lessons still to be learned

• Applicability – Differences in structure and sizes of firms

– Best reporting unit

– Best respondent in reporting unit

• Response Rates• Resources

Page 24: Measuring U.S. Innovative Activities

24

Critical U.S. Data Gaps• Neither an innovation survey nor more

systematic data on innovative activity would fill core data gaps

• Innovative activities of interest as potential drivers of economic outcomes of interest, such as total factor productivity

Page 25: Measuring U.S. Innovative Activities

25

Critical U.S. Data Gaps• Do not cover all of services

– No annual product data for retail & wholesale trade, or services

• Data not designed to be linked– May not be collected as panel

– Links may not work as well as expected

• Small business performance– Need to improve measures, e.g., current SBO vs. 1987 & 1992

CBO

• Data on inputs sparse– Not comparable across sectors

Page 26: Measuring U.S. Innovative Activities

26

Filling Critical Gaps – New Data• FY2008 Budget initiative – quarterly and annual

coverage of all service sectors

• Off shoring & outsourcing indicators – 2007 Economic Census

• Corporate structure– 2007 SBO – add franchising, more characteristics of

business, business owner

– Report of Organization Survey asks imported services in 2007

Page 27: Measuring U.S. Innovative Activities

27

Filling Critical Gaps – New Data• R&D:

– NSF-sponsored addition of questions for selected industries, 2007 Economic Census

– Survey of Industrial Research and Development (SIRD) being redesigned with some of these issues in mind

Page 28: Measuring U.S. Innovative Activities

28

Filling Critical Gaps – Leveraging Existing Data

• More links among micro data sets:– U.S. Patent Data & R&D survey-- NSF-sponsored linkage project

underway

– FDI / DIA & R&D survey -- successful 2005 NSF / BEA / Census pilot

• Expand existing data linkages– Extend M&A database past 1992 -- underway

– LEHD-based analyses – to come

• Link with additional external data– Detailed IT investments – underway

– Financial characteristics – foreign operations of U.S. firms

– R&D tax credit data being linked

– Other administrative data sources -- potential

Page 29: Measuring U.S. Innovative Activities

29

For the Future: Thinking about Measuring Innovative Activities

• Once core gaps are filled

• Begin where we are – e.g. Advisory Committee’s recommendations– Expand NSF’s R&D survey to include broader

range of innovation investment activities– Improve understanding of establishment-firm

relationship– NIPA’s as organizing principle, reinforcing macro-

micro linkages

Page 30: Measuring U.S. Innovative Activities

30

For the Future: Thinking about Measuring Innovative Activities

• Draw on lessons learned– e.g. collecting data on technology, organizational

and management structure: what’s hard, easy– How often?– Surveys vs. censuses

• Plan for longitudinal data– Plan at least 2 rounds

• Need 2, preferably 3, rounds of data to test causality

– Design-in linkages with core data

Page 31: Measuring U.S. Innovative Activities

31

For the Future: Thinking about Measuring Innovative Activities

• Plan for change– 1880: Census added Transportation and Communication

• steam railroads• steamboat companies• incorporated express companies• telegraph companies

– 1999: Information Technology– 2002: Supply Chains– 2005: FDI / DIA– 2009? – 2019?

Page 32: Measuring U.S. Innovative Activities

32

Conclusion

• Understanding role of innovative activity requires – Solid data on core measures of economy– Data on innovative activity– Longitudinal measures of both

• Filling these gaps would improve our ability to understand – Entire economy – Economic performance: TFP, other measures

Page 33: Measuring U.S. Innovative Activities

33

Appendix: The Census Bureau Research Data Center Program

• Purpose

• Research Data Centers (RDCs) offer qualified researchers restricted access to confidential economic and demographic data collected by the Census Bureau in its surveys and censuses. RDCs are established through partnerships with academic and similar organizations.

Page 34: Measuring U.S. Innovative Activities

34

RDC Locations

• Ann Arbor MI

• Berkeley CA

• Boston MA

• Chicago IL

• Ithaca NY

• New York NY

• Los Angeles CA

• Research Triangle NC

• Upper Marlboro MD

• RDCs are Census Bureau facilities staffed by a Census Bureau employee and meet all physical and computer security requirements for restricted access

Page 35: Measuring U.S. Innovative Activities

35

Datasets Available at RDCs

• Most Economic Censuses and Surveys

• Selected Demographic Surveys and Decennial Censuses

• Special Files crated for research purposes:

• Longitudinal Research Database, Integrated Longitudinal Business Database, linked files (e.g., LEHD), etc.

Page 36: Measuring U.S. Innovative Activities

36

RDCs: Title 13 and Title 26

• Our legal authority to provide access to these data require that there be a legitimate Title 13, Chapter 5 purpose to any research.

• Our legal authority to provide access to IRS (Title 26) data in our custody requires that the Title 13 benefit be the predominant purpose.

Page 37: Measuring U.S. Innovative Activities

37

RDCs: Review and Approval Process

• 1. Researcher interacts with an RDC Administrator to gather information about the data available and the procedures to be followed.

• 2. Proposals are submitted for review by Census Bureau staff and are either approved or denied. If denied, they can be resubmitted if they are revised to address noted deficiencies.

Page 38: Measuring U.S. Innovative Activities

38

RDCs: Review and Approval Process, continued

• 3. If approved by the Census Bureau, the proposal must also be approved by the agency “owning” the data (e.g., IRS or the survey sponsor).

• More information:• www.ces.census.gov.