Subsidyscope

Note:  this summary is a joint posting by Doug Koplow (Earth Track) and Ron Steenblik (OECD)

As of the end of June, the Subsidyscope website has been discontinued.  Subsidyscope was an initiative of the Pew Charitable Trusts to examine federal subsidies across multiple sectors of federal activity.  The project, which included significant involvement by the Sunlight Foundation, ran from 2008-2012. 

During 2013, there was no new work other than updating the project's tax expenditure database. The various datasets developed by the project have been archived offline.  Reports and commentary commissioned by Pew remain accessible via Pew's main website, here.

Both of us were on their Advisory Board, along with a number of other people focused on subsidy and budget transparency.  Collectively, the two of us have been working on improving subsidy transparency for several decades at this point.  We thought it might be useful to share some of our takeaways from what Subsidyscope tried to do, the common challenges that these types of initiatives often face, and ways to move subsidy transparency mainstream. 

1) Engage the wonks, make data variance visible 

Specialized knowledge of arcane data systems and data sources was extremely important to bring in early.  Pew did a nice job in this area, engaging people (as staff and advisors) with a broad array of experience in subsidy measurement, data sources, and the politics of subsidy reform.  This group was important not only in identifying potential data sources, but also in understanding the limitations of those data sets and finding people who could help solve the problems that inevitably arose. 

Trying to pull systematic data from the government was a useful exercise for a foundation to take on.  The process highlighted the types of problems that would need to be solved were this type of reporting to become routine, and created an initial map of how standardizing more detailed subsidy reporting might proceed.

The challenges can be fairly sobering.  At one advisory board meeting, participants listed four or five different federal budgeting systems in use, each generating somewhat different spending figures.  It is noteworthy that this disparity applied to the easy subsidies: cash flows being used in a particular budget year.  Support provided through credit or insurance markets, tax expenditures, mandates and regulatory exemptions all require more complicated assessments of activity relative to a hypothetical counter-factual marketplace, and therefore are more difficult to measure. 

Were there a way to easily compare multiple data sources for the same spending area, we expect that the simplest types of data disagreements would quickly be reconciled and that the variance between systems would decline over time. Achieving this type of transparent comparability was one of the main drivers behind Subsidyscope's efforts to integrate tax expenditure data from the Joint Committee on Taxation (JCT) and the Treasury.  Their tables provided a snap-shot not only of differences between expected losses from the same provision across estimators, but also how projected values by the same estimator changed year-to-year.  There are a number of common factors that may result in year-to-year changes, including modifications to eligibility, faster market growth, or changing assumptions on how frequently a tax subsidy is being used.  Rarely are the numerical shifts accompanied with explanations for the changes, however.  Subsidyscope's presentation format made differences in estimated revenue losses easy to see for the first time - making it possible to begin pushing back to root causes.

In an ideal world, this work would have led to changes in behavior and reporting in both the JCT and the Treasury - creating a more visible data set on tax expenditures going forward.  For, as with the budget data variance, unless these comparisons are standardized, routinized, and made accessible to the public, the pressure to identify and correct estimation error (or at least to explain its causes) quickly dissipates. 

Unfortunately, we have not seen a shift in best practices by the tax expenditure estimation teams.  The end of the Subsidyscope initiative will lessen the pressure for better tax expenditure reporting.  At present, no other groups provide the comparative data on tax expenditure that Subsidyscope did.  In truth, though, there is really no reason for reporting not to be improved within the government.  With about $1 trillion/year in federal tax expenditures, is it really too much to ask that the JCT and the Treasury publish their revenue-loss estimates in machine-readable Excel files rather than difficult-to-work-with PDFs?  Or for them to acknowledge, and to routinely report, on how and why their estimates differ from each other, over time, and from actual claims on tax returns?[1]

2)  Some of the most important findings
of Subsidyscope relate to process rather than data sets
 

Looking in detail across multiple sources of government data provided a great opportunity for Pew and Sunlight researchers to identify gaps, discrepancies, and errors in data sources; and to compare and contrast the ease and accuracy with which different parts of the federal government made information available.  In the process of trying to develop work products and reports, these obstacles were frustrating - indeed infuriating - to overcome.  However, when one looks at Subsidyscope as a test of what it takes to bring federal transparency to the next level, looking for patterns in these obstacles and describing how they were (or could be) surpassed becomes vitally important.  

As the project was winding down and people involved with the detailed research were beginning to move to other tasks and opportunities, Doug suggested that the staff do an internal summary of these challenges and lessons while the memories were still fresh.  Our understanding is that such reviews were conducted, and (we hope) shared with Pew's Board. 

These summaries would have been quite interesting for the Advisory Board, and even the public to see as well.  Perhaps Pew will consider making a version public.  A core objective of Subsidyscope was for its work to spark further analysis, disclosure, and reform by other organizations and agencies.  The learning from Pew's four year effort would likely be useful to a broad audience, and perhaps enable some of them to avoid pitfalls as well.

3)  Limits of foundations 

Both Pew and Sunlight are well funded, highly competent organizations.  But trying to build detailed, recurring, core data on government operations is likely beyond the scope of what even large foundations can do.  The work is immense, and the greatest leverage to adjust data collection in ways that streamline informational availability and improve accuracy exists "up-stream" within government agencies - not among users of existing government datasets.

A useful analogy for the path that subsidy reporting needs to follow is corporate reporting before the Securities and Exchange Commission.  At that time, what data there were on corporations and their activities were spotty, inconsistent, and often held by insiders only.  Published reports were unreliable.

Foundations could not have changed that situation.  The cost of initially collecting data is high; but so too is the cost of keeping a data set current year-after-year.  Yet ongoing information over a long period of time is required if people's expectations and decision-making criteria are to be changed.  Subsidyscope is not alone in this regard: many, many NGO data initiatives are funded for a few years, after which their data go stale.

Had Pew funded a test project on corporate reporting in the 1920s for a period of four years, merely demonstrating that better data was logical and welfare-enhancing would not have been sufficient to trigger broad change in reporting.  When corporate reporting did evolve -- through regular mandatory public reporting of financial data for traded companies, independent audits by third parties, and civil and criminal penalties for willful misrepresentation, it was the result of government requirements, not NGO or foundation case studies.  It was a structural change in the way a large group of firms had to operate.

But equally important to remember is that detailed case studies such as what Pew conducted via Subsidyscope can greatly inform the types of regulatory changes needed; and that wide-ranging societal benefits from standardized corporate reporting resulted.  Markets became more transparent, the cost of raising capital fell, and public investors faced a much lower risk (not zero, of course) of being swindled in financial markets.  As such, the system became at least partially self-reinforcing.  Even the firms themselves saw benefits from the new approach; certainly third-party auditors had an interest in the new system surviving.  And so it did.

A similar set of changes are needed in the subsidy-reporting arena.  And, as with corporate reporting, many of the initial components of these changes will be mandatory, not voluntary.  This is a continuation of prior practice, not a departure: though US subsidy reporting is far from perfect or complete, it is salient that some of the biggest steps forward in transparency have been regulatory and legislative in nature.  This includes annual reporting of tax expenditures, audit requirements on government agencies, inspector generals within agencies to root out fraud and abuse, and the tracking and reporting of credit subsidies.

4)  Measuring success 

We'd be interested to hear how Subsidyscope's funders measure the success and failures of their four year effort. Every project has both elements, of course; and both are important to learn from.  But one critical evaluative factor we think needs to be included in projects such as this one is whether the effort directly or indirectly changed the reporting practices of data generators such that transparency would remain improved even after the initiative ended.  Assembling existing data in new ways from the outside is important.  But moving best practices upstream, as well as establishing a recurring process for internal data evaluation and correction, is equally so.

A second critical success factor is whether the work conducted within a specific geographic and temporal setting is able to be linked with, and integrated into, related work on a similar topic conducted by other parties. The ability to integrate learning and information is what enables a narrow set of information to build over time into something broad enough to shift decision making within an economy; and to extend the usefulness of the work beyond the termination point of a  specific project.

Cross-country collaboration on subsidy transparency and reform, including data integration, is ever more achievable.  This is largely the combined result of the internet and the growing number of researchers working in the subsidy area.  We hope to see this potential increasingly leveraged going forward.

For many years, the main route for convergence and standardization of subsidy reporting was through inter-governmental organizations. In the area of agricultural support, for example, the Organisation for Economic Co-operation and Development (OECD) started the ball rolling in the 1980s with its standardized reporting on government support for major crops and livestock commodities.   A similar reporting format was later adopted by the World Trade Organization (WTO) in the mid-1990s. The requirement to notify the WTO about government support to industry led to some convergence among countries in how the data on support were compiled at the national level as well.  However, the weak enforcement mechanism for WTO reporting of industry supports has limited the efficacy of this change lever.

Increasingly, non-governmental organizations such as the Global Subsidies Initiative and the International Budget Partnership  are demonstrating their capability to collect information on subsidies and compile them in a standardized format. Their value-added is clear, though as down-stream users their ability to alter the way data is collected and reported by the government itself is limited. 

Subsidyscope's focus was on the United States.  But additional outreach to similar initiatives in other countries could well have leveraged their domestic work - both by providing mutual support on common challenges and by helping to achieve improved coherence between the US results and work conducted abroad.

5)  Rescuing data from archival oblivion 

Project wind-down normally involves a good deal of filing and archiving.  Data records are "preserved" on computer drives, but in terms of policy usefulness, they are highly perishable commodities.  Older data become less relevant to current policy decisions, but if integrated into a time series or comparative metrics, can remain important for an extended period of time. 

Yet once archived, knowledge about the data sets is rapidly lost.  Fixes done to make them workable, sources of raw data, programming routines to integrate or normalize information from multiple sources, even definitions of field names get misplaced.  Much of the meta data on these data sets, so critical in extending their use, resides in the heads of the original research team.  As these people move on to new positions, this important information disperses with them.  Data storage protocols are always changing as well, and electronic data from 15 years ago may be difficult even to read into a computer today; data from 30 years ago are nearly irrretrievable.

Again, these are not problems unique to the Subsidyscope effort, by any means.  The Open Knowledge Foundation focuses on the issue of "data salvage" and reuse straight out, and across a variety of research fields.  But the question we pose is whether there is a way to archive subsidy data differently going forward - so the work by Subsidyscope, or by environmental NGOs, or indeed by analysts within the governments generating the raw data themselves, can be more readily integrated as a module into a bigger data "whole" rather than sent to data oblivion.  Ideally, these disparate pieces could be structured in such a way so that they can be  easily integrated into a larger dataset, and thereby contribute to a broadening picture of government subsidies.

 


[1] In a 2007 paper Doug prepared for the OECD, he advocated requiring a formal variance report each year by JCT and Treasury any time their estimates were more than $100 million off from what was actually claimed on tax returns.  A similar report should be required whenever JCT and Treasury estimates for the same provision differ by more than $50 million.  The goal isn't to have perfect accuracy in estimation, but to establish a process by which errors are better understood and the cost of tax breaks can be estimated with greater precision over time.

 

Natural gas fracking well in Louisiana

Improved visibility on tax expenditures was one of my biggest hopes for Pew's Subsidyscope when I joined the advisory board, so I am very happy to see the first step of this launched.  Have a look at it here.  With an aggregate value of more than $1 trillion in 2009 (see page 12), it is high time to focus more attention and resources to establish better transparency and accountability in this area. 

Yes, I know the arguments about tax breaks:  "What's the matter with just letting people keep more of their own money?"  Nothing, of course, if we take the question on a metaphysical level.  But the real world is one where all of us use government services, and reductions in taxes to one group must be made up by higher taxes on another.  As a result,  who gets the special deals, and how much cash they save through those special deals, matter a whole lot both in terms of fairness and for the long-term competitiveness of our economy.  After all, the biggest subsidies of any form generally go to the most politically powerful existing industries, not to the as-yet-undefined, struggling to survive start-ups on which the long-term growth of our economy often relies. 

Estimating tax expenditures is not an easy task, and both JCT and Treasury have relatively small staffs to do it.  With revenue losses so high though, this is not a valid excuse.  Maybe we need more people, better data systems, and unique tax break identification numbers so we can more easily track usage of each tax break so we can do the job better.  Maybe large tax breaks need to be analyzed with the same rigor as we require for new regulations.

What Pew has done here is admittedly a relatively small step.  They have centralized data on tax breaks done by two federal institutions, the US Treasury and the Joint Committee on Taxation, in one place, allowing citizens to easily compare estimates for the same line item for the first time.  This should have been done years ago by the institutions themselves, rather than requiring investments of Foundation money.  But even this small step is a very important one.  It allows questions to be easily asked that were more difficult to ask before.  And asking those questions can help to move us along in the proper direction of treating our investments in tax reductions to selected groups as real money that needs to have real oversight and to be killed if better returns for taxpayers are available through other means. 

Where JCT and Treasury numbers in the Pew database are very different, it would be helpful for the respective estimation groups to jointly produce a one page summary on why.  This can be called a "Tax Expenditure Variance Report," and should be mandated by Congress any time the annual estimates differ from each other by $100 million or more. 

Other changes I think need to happen in this area:

  • Both Treasury and JCT should, as a matter of course, issue their estimates in electronic formats that can be easily imported into software programs such as Excel for further evaluation.  These data sets should be populated as far back as the two organizations have data; and should adjust for name changes where the basic subsidy has remained to allow for a proper time series.
  • Tax expenditure estimates that cut across multiple sectors (accelerated depreciation, for example) should be broken out by industry sector.  The rules are not always identical in each one, and it is important to be able to gauge the relative distribution of these benefits.
  • Meta data on institutional structures that bypass corporate taxation entirely should be distilled into summary tables that a general user can understand.  This would help to capture the differential use of pass-through vehicles (LLCs, PTPs, MLPs, etc.) and corporate inversions across industrial sectors that can generate quite different tax burdens on different ways to supply similar goods and services. 
  • Meta data on average effective tax rates by industry should also be produced annually by government.  Where this has been generated in the past, it has usually been by NGOs using far more limited information on taxes paid than what the government itself posesses.
  • A formal process to compare forward-looking estimates with actual losses incurred under each tax expenditure should be instituted, again with publicly-available variance reports that describe inaccuracies that exceed $100 million per line item.
  • While the organizations don't want to release their estimation models to prevent gaming by industry and legislators, some use of outside review should be instituted, especially if the variance between projected revenue losses and actual are quite large.
  • Pew and others need to continue to expand their coverage to be sure that sectoral comparisons of aggregate tax breaks are accurate.  For example, the current version of the database excludes excise tax breaks, which for ethanol are approaching $6 billion per year and would generate a big jump in the aggregate tax breaks for energy.  Tracking of tax breaks should be as comprehensive as possible over time if it is to inform decision making in a neutral way.
Natural gas fracking well in Louisiana

Congratulations to the Marcus Peacock's team at Subsidyscope (and, to Kevin Webb's group at the Sunlight Foundation -- they have also played an instrumental role), for the launch of their energy subsidy module yesterday.

The energy sector is the fourth module to be completed by Subsidyscope.  Earlier releases have included a review of subsidies in the financial bailout, transport subsidies, and subsidies to non-profit institutions.  Support for housing, and cross-cutting work on integrating data on tax expenditures are currently underway.

Because government involvement with energy cuts across so many agencies and activities, tracking energy subsidies has been a challenging task.  Inevitably, not everything could be addressed in the first go, and some of the issues they hope to address in the coming months are summarized here.  As programs are added, expect susidy numbers to rise substantially.  Better resolution on accelerated depreciation benefits, impacts of purchase mandates such as the Renewable Fuel Standards, and excise tax exemptions (they've focused on income taxes initially) will each add billions of dollars per year.SSegy

In my role as an external advisor to this project over the past few years, I've been struck by a couple of things about Pew's engagement with the subsidy issue.  First, they have been willing to tackle a set of very complicated policy areas recognizing full well that complete exposition of sector subsidies would need to be built gradually over time.  This iterative approach is something that the G20 fossil fuel subsidy reform effort will likely need to embrace as well. 

Second, due to their scale and contacts, Pew has been able to engage directly with key officials in relevant government agencies in order to expand the available information set on subsidy programs.   This is quite important, and something that I hope will succeed in permamently changing the way some of this information is compiled and released.  Over the long-term, more timely, comprehensive, and comparable data sets in areas such as federal tax expenditures, credit, and insurance subsidies alone would have immense value in improved governance and economic efficiency.