This is the accessible text file for GAO report number GAO-08-440 entitled 'Chemical Assessments: Low Productivity and New Interagency Review Process Limit the Usefulness and Credibility of EPA's Integrated Risk Information System' which was released on April 29, 2008. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to the Chairman, Committee on Environment and Public Works, U.S. Senate: United States Government Accountability Office: GAO: March 2008: Chemical Assessments: Low Productivity and New Interagency Review Process Limit the Usefulness and Credibility of EPA's Integrated Risk Information System: GAO-08-440: GAO Highlights: Highlights of GAO-08-440, a report to the Chairman, Committee on Environment and Public Works, U.S. Senate. Why GAO Did This Study: The Environmental Protection Agency’s (EPA) Integrated Risk Information System (IRIS) contains EPA’s scientific position on the potential human health effects of exposure to more than 540 chemicals. IRIS is a critical component of EPA’s capacity to support scientifically sound environmental decisions, policies, and regulations. GAO was asked to examine (1) the outcome of steps EPA has taken to ensure that IRIS contains current, credible chemical risk information, to address the backlog of ongoing assessments, and to respond to new requirements from the Office of Management and Budget (OMB); and (2) the potential effects of planned changes to the IRIS assessment process on EPA’s ability to ensure that IRIS provides current, credible risk information. To do this work, GAO reviewed and analyzed EPA data and interviewed officials at relevant agencies. What GAO Found: EPA’s actions since 2000 to ensure that IRIS contains current, credible risk information, to address its backlog of 70 ongoing assessments, and to respond to new OMB requirements—including increasing funding and revising the assessment process—have not enabled EPA to routinely complete credible IRIS assessments or decrease its backlog. Although in fiscal years 2006 and 2007 EPA sent 32 assessments to OMB for the first of three required external reviews, EPA finalized only 4 assessments during this period. This low level of productivity jeopardizes the viability of the IRIS database. Further, an EPA analysis indicated that many existing assessments may need to be updated, and EPA program offices and other IRIS users have requested assessments of hundreds of chemicals not yet in IRIS. Factors contributing to EPA’s inability to complete IRIS assessments in a timely manner include new OMB-required reviews of IRIS assessments by OMB and other federal agencies; certain EPA management decisions, such as delaying some assessments to await new research; and the compounding effect of delays—even one delay can have a domino effect, requiring the process to essentially be repeated to incorporate changing science. As of December 2007, most of the 70 ongoing assessments had been in progress for over 5 years. Regarding new OMB requirements, the IRIS assessment process now includes two OMB/interagency reviews of draft assessments. These reviews have resulted in involvement of other federal agencies in EPA’s IRIS assessment process in a manner that limits the credibility of IRIS assessments and hinders EPA’s ability to manage them. That is, the OMB/interagency reviews lack transparency—OMB considers agencies’ comments on IRIS assessments to be internal executive branch documents that may not be made public. Given the importance of IRIS assessments, it is essential that input from all parties, including other federal agencies, be part of the public record. Transparency is especially important because agencies providing input include those that may be affected by the assessments should they lead to regulatory or other actions. Also, without communicating its rationale for doing so, OMB directed EPA to terminate five assessments that for the first time addressed acute, rather than chronic, exposure—even though EPA initiated this type of assessment to help it implement the Clean Air Act. Most OMB/interagency reviews completed to date have added 6 or more months to the IRIS time frames. Such delays and credibility concerns would likely be exacerbated by further changes EPA is planning to respond to continuing concerns of other federal agencies, such as providing them with an expanded role in EPA’s IRIS assessment process and discretion to suspend assessments to develop new studies for some chemicals. EPA estimates that such assessments would take up to 6 years, an estimate GAO believes is conservative in light of the assessment time frames under the current process. Suspending assessments is inefficient; alternatively, with longer-term planning, EPA could provide agencies and the public with more advance notice of assessments, enabling them to complete relevant research before IRIS assessments are started. What GAO Recommends: GAO recommends that EPA (1) clearly define and document an IRIS assessment process that, among other things, can be conducted within a time frame that minimizes the need for rework and (2) ensure that it can develop transparent, credible assessments by, for example, determining the types of IRIS assessments it will conduct based on EPA program needs and defining the appropriate role of other federal agencies in its IRIS assessment process. EPA agreed to consider GAO’s recommendations in revising the IRIS assessment process. To view the full product, including the scope and methodology, click on [hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-08-440]. For more information, contact John B. Stephenson at (202) 512-3841 or stephensonj@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: EPA's Efforts to Improve the IRIS Assessment Program Have Not Produced the Desired Results: While Appropriate Coordination with Federal Agencies Could Help EPA Resolve IRIS Assessment Controversies More Efficiently, EPA's Proposed Expansion of Agencies' Roles in IRIS Assessments Would Cause Further Delays and Limit Their Credibility: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendix I: Objectives, Scope, and Methodology: Appendix II: Broad Outline of EPA's Current IRIS Assessment Process: Appendix III: Information on OMB/Interagency Reviews of IRIS Assessments: Appendix IV: Comments from the Environmental Protection Agency: Appendix V: Comments from the Office of Management and Budget: Appendix VI: GAO Contact and Staff Acknowledgments: Figures: Figure 1: National Academies' Risk Assessment and Risk Management Model Used by EPA: Figure 2: Funding for the IRIS Program, Fiscal Years 2000-2007: Figure 3: Number of Completed IRIS Assessments, Draft Assessments to OMB, and IRIS Staff in Full-Time Equivalents, Fiscal Years 2000-2007: Figure 4: December 2007 Status of IRIS Assessments That EPA Reported as Ongoing in Fiscal Year 2007: Figure 5: EPA's Draft Proposed IRIS Assessment Process: Figure 6: Status of IRIS Assessments Sent to OMB for the First OMB/ Interagency Review Starting in Fiscal Year 2004, as of December 1, 2007: Abbreviations: ATSDR: Agency for Toxic Substances and Disease Registry: DOD: Department of Defense: DOE: Department of Energy: EPA: Environmental Protection Agency: IARC: International Agency for Research on Cancer: IRIS: Integrated Risk Information System: NASA: National Aeronautics and Space Administration: NCEA: National Center for Environmental Assessment: NCI: National Cancer Institute: NIOSH: National Institute of Occupational Safety and Health: OMB: Office of Management and Budget: ORD: Office of Research and Development: PART: Program Assessment Rating Tool: PBPK: physiologically based pharmacokinetic: RfC: inhalation reference concentration: RfD: oral reference dose: TCE: trichloroethylene: [End of section] United States Government Accountability Office: Washington, DC 20548: March 7, 2008: The Honorable Barbara Boxer: Chairman: Committee on Environment and Public Works: United States Senate: Dear Madam Chairman: The Environmental Protection Agency's (EPA) Integrated Risk Information System (IRIS)--a database integral to the agency's mission of protecting human health and the environment--contains EPA's scientific position on the potential human health effects that may result from exposure to various chemicals in the environment. IRIS data provide the fundamental scientific components needed to develop human health risk assessments. These health risk assessments, in turn, provide the foundation for risk management decisions, such as whether EPA should establish air and water quality standards to protect the public from exposure to toxic chemicals or set cleanup standards for hazardous waste sites. In addition, state and local environmental programs, as well as some international regulatory bodies, rely on IRIS health effects information in managing their environmental protection programs. Although the information in IRIS is a critical primary component of EPA's capacity to support scientifically sound decisions, policies, and regulations, many IRIS assessments are outdated, and few assessments have been completed in recent years. This has resulted in a significant backlog of incomplete chemical assessments and a growing number of outdated assessments. Further, while EPA's IRIS database currently includes about 540 chemicals, every year approximately 700 new chemicals enter commerce, any number of which could pose significant human health risks. Overall, the goal of the IRIS assessment process is to produce quantitative estimates of cancer and noncancer effects from chronic (long term) exposure to the chemicals assessed. One impact of not having current and complete IRIS assessments of many potentially harmful chemicals is that some chemicals that pose health risks to the public may not be regulated under, for example, air or drinking water statutes, or are regulated by standards that may not sufficiently take into account the best available science on human health effects. For example, trichloroethylene (TCE), a solvent widely used as a degreasing agent in industrial and manufacturing settings, is the most frequently reported organic contaminant in groundwater and has been linked to cancer and other health hazards, according to the National Academies.[Footnote 1] Yet, because of questions raised by peer reviewers about the IRIS cancer assessment for TCE, EPA withdrew the assessment from IRIS in 1989, did not initiate a new TCE assessment until 1998, and likely will not complete that assessment until 2010 or later. This delay represents an information gap of at least 21 years. Without completed IRIS assessments reflecting current risk data, EPA lacks assurance that its regulatory decisions concerning this widespread chemical reflect the best available science on its potential health effects. While the IRIS assessment process includes numerous individual steps or activities, major assessment steps include (1) a review of the scientific literature; (2) preparation of a draft IRIS assessment; (3) internal EPA reviews of draft assessments; (4) two Office of Management and Budget (OMB)/interagency reviews, managed by OMB that provide for input from OMB as well as from other federal agencies, including those that may be affected by the IRIS assessments if they lead to regulatory or other actions; (5) an independent peer review conducted by a panel of experts; and (6) the completion of a final assessment that is posted to the IRIS Web site. EPA's assessment process has undergone a number of formal and informal changes during the past several years. The agency is planning further changes--particularly in the areas of external reviews and scientific data gaps--largely to address concerns of other federal agencies, such as the Department of Defense (DOD). Some of the assessment process changes have raised concerns about EPA's ability to keep its scientific assessments separate from its risk management decisions, as the National Academies recommends and EPA policy endorses. In this context, this report examines (1) the outcome of steps EPA has taken to ensure that IRIS contains current, credible chemical risk information, to address the backlog of IRIS assessments, and to respond to new requirements from OMB; and (2) the potential effects of EPA's planned changes to the IRIS assessment process on EPA's ability to ensure that IRIS provides current, credible risk information. In conducting our work, we obtained and analyzed information on EPA's productivity, including the number of new and completed IRIS assessments, for fiscal years 2000 through 2007; the status of IRIS assessments, as of December 1, 2007, that were in progress during fiscal year 2007; the status of IRIS assessments that have been sent to OMB for OMB/interagency review; the number of assessments in the IRIS database that may need to be updated; the resources provided to the program for fiscal years 2000 through 2007; and user needs and EPA's assessment completion goals. We interviewed EPA's National Center for Environmental Assessment officials who manage the IRIS assessment program, as well as officials from other EPA program offices and federal science and health agencies that are involved in the IRIS assessment process, to obtain their perspectives on, among other things, the current IRIS assessment process, the potential effects of the proposed changes to the process, the extent to which EPA has made progress in completing assessments and meeting user needs, and challenges EPA faces in completing assessments. In addition, we also interviewed officials from the Department of Defense, the Department of Energy (DOE), the National Aeronautics and Space Administration (NASA), and OMB to obtain their perspectives on the OMB/interagency review process and on the planned changes to the IRIS assessment process. We did not evaluate the scientific content or quality of IRIS assessments. (See app. I for a more detailed description of the methodology we employed.) We conducted our work from October 2006 to March 2008 in accordance with generally accepted government auditing standards, which require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. Results in Brief: EPA has taken a number of steps to help ensure that IRIS contains current, credible chemical risk information, to address its backlog of ongoing assessments, and to respond to new OMB requirements. However, to date, these changes--including increasing funding, centralizing staff conducting assessments, and revising the assessment process--have not enabled EPA to routinely complete credible IRIS assessments or decrease the backlog. That is, although EPA sent 32 draft assessments for external review in fiscal years 2006 and 2007, the agency finalized only 4 IRIS assessments during this time. Several key factors have contributed to EPA's inability to achieve a level of productivity that is needed to sustain the IRIS program and database: new OMB-required reviews of IRIS assessments by OMB and other federal agencies; the growing complexity and scope of risk assessments; certain EPA management decisions and issues, including delaying completion of some assessments to await new research or to develop enhanced analyses of uncertainty in the assessments; and the compounding effect of delays. Regarding the last factor, even a single delay in the assessment process can eventually lead to the need to repeat the assessment process to take into account changes in science and methodologies. A variety of delays have impacted the majority of the 70 assessments being conducted as of December 2007--causing them to be in process for more than 5 years. These time frames are problematic because of the substantial rework such cases often require to take into account changing science and methodologies before they can be completed. Further, because EPA staff time continues to be dedicated to completing these assessments, EPA's ability to both keep the more than 540 existing assessments up to date and initiate new assessments is limited. Importantly, EPA program offices and state and local entities have requested assessments of hundreds of chemicals not yet in IRIS, and EPA data as of 2003 indicated that the assessments of 287 chemicals in the database may be outdated--that is, new information could change the risk estimates currently in IRIS or enable EPA to develop additional risk estimates for chemicals in the database. In addition, because EPA's 2003 data are now more than 4 years old, it is likely that more assessments may be outdated now. One of the factors that has contributed to EPA's inability to complete assessments in a timely manner--the new OMB/interagency review process- -also limits the credibility of the assessments because it lacks transparency.[Footnote 2] Specifically, neither the comments nor the changes EPA makes to the scientific IRIS assessments in response to the comments made by OMB and other federal agencies, including those whose workload and resource levels could be affected by the assessments, are disclosed. According to OMB, the comments it provides to EPA and EPA's disposition of them are considered internal executive branch communications that may not be made public. Further, OMB has not communicated its rationale for directing EPA to discontinue work on five IRIS assessments that EPA had sent to OMB for OMB/interagency review. These assessments, initiated to meet EPA program needs, were the first EPA IRIS assessments of short-term (acute) risks of exposure; the IRIS program historically has evaluated long-term (chronic) risks. The additional assessment process changes EPA is planning would likely exacerbate delays in completing IRIS assessments and further affect their credibility. Specifically, despite the informal OMB/interagency review process that OMB required EPA to incorporate into the IRIS assessment process in 2005, certain federal agencies continue to believe they should have greater and more formal roles in EPA's development of IRIS assessments. Consequently, EPA has been working for several years to establish a formal IRIS assessment process that would expand the role of federal agencies in the process--including agencies such as DOD, which could be affected by the outcome of IRIS assessments. Some of these agencies and their contractors could face increased cleanup costs and other legal liabilities if EPA issued an IRIS assessment for a chemical that resulted in a decision to regulate the chemical to protect the public. Under EPA's planned changes, these potentially affected agencies would be able, at several points in the assessment process, to subject particular chemicals of interest to additional process steps. EPA estimates that these assessments would take up to 6 years to complete because, for example, at the discretion of these agencies, EPA would suspend the assessment process for up to 18 months so the agencies could conduct additional research to fill data gaps, rather than proceeding with currently available data. While it is important to ensure that assessments consider the best science, EPA has acknowledged that waiting for new data can result in substantial harm to human health, safety, and the environment. Further, although coordination with other federal agencies about IRIS assessments could enhance the quality of the assessments, increasing the role of agencies that may be affected by IRIS assessments in the process itself reduces the credibility of the assessments if that expanded role is not transparent. In this regard, while EPA planned to include federal agencies' comments in the public record, the process changes have been delayed since early 2007 in part because of OMB's view that agencies' comments about IRIS assessments represent internal executive branch communications that may not be made public--a view that is inconsistent with the principle of sound science that relies on, among other things, transparency. We are making recommendations to the EPA Administrator to require the Office of Research and Development to re-evaluate its draft proposed changes to the IRIS assessment process in light of the issues raised in this report and ensure that any revised process, among other things, clearly defines and documents an IRIS assessment process that will enable the agency to develop the timely chemical risk information it needs to effectively conduct its mission. In addition, we are recommending that the EPA Administrator take steps to better ensure that EPA has the ability to develop transparent, credible IRIS assessments, including determining the types of assessments it needs to support its programs; defining the appropriate role of external federal agencies in EPA's IRIS assessment process; and managing an interagency review process in a manner that enhances the quality, transparency, timeliness, and credibility of IRIS assessments. We provided EPA and OMB with a draft of this report for review and comment. EPA agreed to consider our recommendations in revising the IRIS assessment process. However, EPA stated that it believed the productivity and transparency issues discussed in the draft report were misrepresented in the title and body of the report. We disagree and believe we have fairly represented IRIS productivity and transparency issues related to the IRIS assessment process. We did, however, clarify that the transparency issues highlighted in our report focus on the IRIS assessment process rather than on the content of IRIS assessments, and we revised the report title. In its comments, OMB did not specifically address the recommendations we made to EPA but disagreed with some aspects of the report, such as our characterization of the purpose and effect of the OMB-managed interagency reviews of IRIS assessments and our conclusion that interagency comments should be transparent. We disagree with OMB and believe that we have fairly represented the OMB/interagency review process as well as the importance of making input from all parties publicly available to alleviate concerns of potential bias. EPA's and OMB's letters and our detailed responses are discussed further at the end of this report and in appendixes IV and V. Background: EPA's Integrated Risk Information System (IRIS) is an important source of information on health effects that may result from exposure to chemicals in the environment. IRIS was created in 1985 to help EPA develop consensus opinions within the agency about the health effects from chronic exposure to chemicals, and its importance has increased over time as EPA program offices and the states have increasingly relied on IRIS information in making environmental protection decisions. Today, the IRIS database--which currently contains assessments of more than 540 chemicals--is heavily relied upon by EPA, state and local environmental programs, international regulatory bodies, academia, industry, and others to support risk-based decision making to protect public health and the environment. According to EPA, national and international users access the IRIS database approximately 9 million times a year. EPA's Assistant Administrator for the Office of Research and Development has described IRIS as the premier national and international source for qualitative and quantitative chemical risk information; other federal agencies have noted that IRIS data is widely accepted by all levels of government across the country for application of public health policy, providing benefits such as uniform, standardized methods for toxicology testing and risk assessment, as well as uniform toxicity values. Similarly, a private-sector risk assessment expert has stated that the IRIS database has become the most important source of regulatory toxicity values for use across EPA's programs and is also widely used across state programs and internationally. As shown in figure 1, the toxicity assessments in the IRIS database fulfill the first two critical steps of the risk assessment process-- providing hazard identification and dose-response assessment. IRIS information can then be used with the results of exposure assessments (typically conducted by EPA's program or regional offices) to provide an overall characterization of the public health risks for a given chemical in a given situation. The risk characterization information can be used to make risk management decisions designed to protect public health. The development of risk assessments is thus directly dependent on the development of toxicity assessments such as those developed in the IRIS program. Figure 1: National Academies' Risk Assessment and Risk Management Model Used by EPA: [See PDF for image] This figure is an illustration of the National Academies' Risk Assessment and Risk Management Model Used by EPA. The following information is depicted: Risk assessment: * IRIS toxicity assessment: 1. Hazard identification; 2. Dose-response assessments. * Exposure assessment; * Together these assessments form: Risk characterization. Risk management: * Development of regulatory options; * Evaluation of public health, economic, social, and political consequences of regulatory options; * Added together with risk characterization, these yield: * Agency decisions and actions. Source: National Academies. {End of figure] Risk management, as opposed to risk assessment, involves integrating the risk characterization information generated from the risk assessment with other information, such as economic information on the costs and benefits of mitigating the risk, technological information on the feasibility of managing the risk, and the concerns of various stakeholders, to decide how to protect public health. An initial risk management decision would be to determine whether the health risks identified in a chemical risk assessment warrant regulatory or other actions. Examples of subsequent risk management decisions that could stem from a determination that action is necessary to protect public health include deciding (1) how much of a chemical a company may discharge into a river; (2) which substances may be stored at a hazardous waste disposal facility; (3) the extent to which a hazardous waste site must be cleaned up; (4) permit conditions for treatment, storage, or disposal of hazardous waste; (5) levels for air emissions; and (6) allowable levels of contamination in drinking water. Thus, as EPA has recognized, although IRIS assessments are not regulatory in nature, the quantitative IRIS values may influence many environmental decisions and may serve as a basis for regulatory consideration. A typical IRIS assessment contains a qualitative hazard identification description and quantitative dose-response assessments. Among other things, a hazard identification description identifies the potential noncancer and cancer health effects of exposure to a chemical that research studies have suggested or determined. For example, for cancer effects, EPA describes the potential health risk using one of five weight-of-the-scientific-evidence descriptors, ranging from "carcinogenic to humans" to "not likely to be carcinogenic to humans." The quantitative assessments also address noncancer and cancer health effects and are developed if there are sufficient credible research data, primarily from either animal (toxicity) or human (epidemiology) studies to support this type of analysis. The noncancer dose-response assessments may include: * an oral reference dose (RfD)--an estimate of the daily exposure to a chemical that is likely to be without an appreciable risk of deleterious effects during a person's lifetime--expressed in terms of milligrams per kilogram, and: * an inhalation reference concentration (RfC)--an estimate of the daily exposure to a chemical that is likely to be without an appreciable risk of deleterious effects during a person's lifetime--expressed in terms of milligrams per cubic meter. The quantitative cancer toxicity assessments include estimates of a chemical's carcinogenic potency--a "cancer slope factor" and "unit risks." Both the cancer slope factor and unit risks are estimates of the increased cancer risk (e.g., 1 in 100 people getting cancer, 1 in 1,000, 1 in 1 million, etc.) from a lifetime of exposure to a given chemical. However, the unit risk is an estimate of the increased risk for lifetime exposure at a standard concentration of a chemical in air or water (1 microgram per cubic meter of air or 1 microgram per liter in water), whereas the cancer slope factor is an estimate of the increased risk per unit dose (calculated using a dose-response curve, or a graph that shows the relationship between a dose and the proportion of exposed persons or animals that have a biologically significant response).[Footnote 3] Historically and currently, the focus of IRIS toxicity assessments has been on the potential health effects of long-term (chronic) exposure to chemicals. According to OMB,[Footnote 4] EPA is the only federal agency that develops qualitative and quantitative assessments of both cancer and noncancer risks of exposure to chemicals, and EPA does so largely under the IRIS program. Other federal agencies develop qualitative cancer assessments or quantitative estimates of noncancer effects of exposure to chemicals in the environment. For example, the Agency for Toxic Substances and Disease Registry (ATSDR) develops quantitative estimates of the noncancer effects of exposures to chemicals in the environment for exposures of up to 14 days (acute); more than 14 days but less than a year (subchronic); and 365 days and longer (chronic). While ATSDR's toxicological profiles include information from other agencies' cancer assessments, including EPA's quantitative IRIS cancer assessments, ATSDR does not develop quantitative cancer assessments. ATSDR toxicological profiles also include qualitative cancer assessments developed by the Department of Health and Human Services' National Toxicology Program and the World Health Organization's International Agency for Research on Cancer (IARC). While these latter assessments provide information on the effects of long-term exposures to chemicals, they only provide qualitative assessments of cancer risks (e.g. known human carcinogen, likely human carcinogen, etc.) and not quantitative estimates of cancer potency, which are required to conduct quantitative risk assessments. As the IRIS database became more widely used and accepted, EPA took steps, beginning in the early 1990s, to improve and maintain the IRIS program and database. For example, the agency created an IRIS Quality Action Team that produced a report in 1994 outlining a number of recommendations for improvement, and EPA implemented a pilot program from 1995-1997 to test new operational procedures for the IRIS program. Changes under the pilot program included developing a standard toxicological review document to support each IRIS summary, incorporating peer review into the assessment process, and establishing a standing group of 18 senior health scientists to conduct the internal agency review of all IRIS draft assessments. The standing group, now called the IRIS Agency Review Committee, includes representatives from the program offices and regions. EPA also formed an IRIS implementation strategy team that developed recommendations in 1997 to improve the IRIS program and database. Key recommendations addressed the need to (1) update IRIS information, (2) establish an annual agenda for the program, (3) form a central IRIS staff to be responsible for the database and to coordinate with the program and regional offices leading individual IRIS assessments, (4) provide Internet access to the IRIS database, and (5) conduct more outreach to users. In response, EPA formed a small centralized IRIS staff in the Office of Research and Development, National Center for Environmental Assessment (NCEA), which has implemented many of these recommendations. For example, EPA placed the IRIS database on its Web page, set up a hot line service to improve outreach with users, and started developing an annual IRIS agenda that identifies the chemicals to be assessed during the fiscal year (new and ongoing assessments) and providing this agenda to the public in a notice in the Federal Register. EPA also responded to the recommendations by posting external peer review drafts of IRIS assessments on EPA's Web page and considering public comments received on these drafts. As discussed below, EPA has continued to evaluate its IRIS assessment process and make changes in an effort to improve it. EPA's Efforts to Improve the IRIS Assessment Program Have Not Produced the Desired Results: In response to criticisms and in an effort to meet the needs of the programs and entities that rely on IRIS information, EPA has continued to make a number of IRIS assessment process changes aimed at improving the timeliness, quality, consistency, and transparency of IRIS assessments. However, even with process changes and increased program funding and staffing, EPA has not been able to routinely complete credible assessments or decrease its backlog of ongoing assessments. Several key factors have contributed to EPA's inability to achieve a level of productivity that is needed to sustain the IRIS program and database, including the OMB/interagency review process managed by OMB, certain management decisions and issues regarding the IRIS program, and the compounding effect of delays. In addition, because the OMB/ interagency review process is not transparent, this change also limits the credibility of IRIS assessments. EPA Efforts to Improve IRIS Continue: Despite the many steps that EPA took throughout the 1990s aimed at improving the IRIS assessment process, the agency has continued to face criticism that the risk information in the database is outdated and of varying quality. For example, according to congressional testimony in 2000 by a risk assessment expert and a representative of a chemical industry association, the outdated and inconsistent information in IRIS represented a serious limitation that undermined the accuracy of risk assessments and risk management decisions. In addition, external parties--notably entities that may be affected by the IRIS assessments, including other federal agencies and industry--have criticized the IRIS assessment process as lacking transparency and have sought earlier input into EPA's assessment process. Further, as a result of continuing concerns that EPA and state regulators were relying on scientific information that was potentially outdated, in 2000, a Senate appropriations committee report directed EPA to conduct an assessment to determine the need to both update and add new assessments to IRIS. In response to the criticisms and in an effort to meet the needs of the programs and entities that rely on IRIS information, EPA has continued to make a number of IRIS assessment process changes aimed at improving the timeliness, quality, consistency, and transparency of IRIS assessments. For example, EPA further centralized the assessment process, hiring additional scientists in the Office of Research and Development to lead individual IRIS assessments; in the past, chemical managers in various program offices and regions had led the assessments on a voluntary basis. EPA made this change to improve the timeliness of assessments and to address concerns about the inconsistency of assessments that the different offices were producing. In addition, EPA changed its peer review requirements, opting for peer review panels, whose meetings the agency has opened to the public, rather than obtaining written peer review comments from several experts. According to EPA, these steps were taken to provide the best possible scientific review of each assessment and to make the review process more transparent. EPA also decided to conduct its internal agency review of assessment drafts earlier in the process. That is, the 18-member IRIS Agency Review Committee comprising EPA senior health scientists now comments on draft assessments before they are released for external review, as well as after external review comments are incorporated in the assessments. In the past, the review by the IRIS Agency Review Committee had taken place once, following the external reviews. According to EPA, this change was made to enhance peer reviews by identifying key science issues and providing external reviewers with drafts that had already been thoroughly vetted within EPA. Prior to this change, IRIS assessments had been internally peer reviewed by three or four scientists with relevant expertise in the Office of Research and Development, program offices, or regions before the assessments were sent to external peer review.[Footnote 5] EPA has also added formal briefings of draft assessments to the Assistant Administrator, Office of Research and Development, at various stages in the assessment process. In addition, EPA informally elevated final IRIS assessment approval authority to the Assistant Administrator, Office of Research and Development. EPA has also delayed or suspended some assessments to await new or scientific studies. Another key change that EPA has incorporated into the IRIS assessment process at OMB's request is an OMB/interagency review process that is managed by OMB. The purpose of these reviews is to obtain input from OMB and other federal agencies that OMB has determined have an interest in particular IRIS assessments as they are being developed in order to help ensure and increase their quality. The reviews occur at two points in the IRIS assessment process--first, after the internal agency review but before the external peer review; and second, after EPA has incorporated input from the external peer review. According to EPA officials, this OMB/interagency review process has evolved from an ad hoc review of selected IRIS assessments of interest to OMB and other federal agencies to a process that now requires, for all assessments, OMB's determination that EPA has satisfactorily addressed all OMB/ interagency comments. This determination must be made both before EPA can provide draft assessments to external peer reviews and before EPA can finalize and post assessments on the IRIS database. Although the IRIS assessment process continues to evolve, it currently typically includes the following steps (see app. II for a flow chart of the process): * solicitation of chemical nominations from EPA program and regional offices, federal agencies, and the public. * selection of chemicals to be assessed during the fiscal year (referred to as the annual agenda). * publication of a Federal Register notice announcing EPA's annual agenda--the specific chemical assessments the agency intends to conduct--and requesting scientific information about these chemicals, thereby giving the public and other federal agencies an opportunity to identify relevant studies. * a review of the scientific literature. * development of draft IRIS toxicological reviews and summaries containing qualitative and, if sufficient information is available, quantitative risk estimates, that have undergone internal peer consultation--a peer review by three or four EPA scientists with relevant expertise. * initial internal agency review, which includes review and comment from the 18-member EPA IRIS Agency Review Committee. * OMB/interagency review by other federal agencies, such as DOD, DOE, and NASA, coordinated by the Office of Management and Budget; OMB informs EPA when EPA has adequately addressed interagency comments. * external peer review by a group of independent experts, convened by an EPA contractor, the Science Advisory Board, or the National Academies, and public comment. * a second internal agency review. * a second OMB/interagency review; OMB informs EPA when EPA has adequately addressed interagency comments. * EPA management review and approval. * posting of IRIS assessments on the IRIS database, available on EPA's Web site. In addition to EPA continuing to revise its IRIS assessment process, since fiscal year 2000, funding for the IRIS program has increased-- from $1.7 million to $9.6 million in fiscal year 2007 (see fig. 2). The need for increased resources to accomplish significant IRIS improvements had been noted in the February 1997 IRIS Implementation Strategy Team Report. Figure 2: Funding for the IRIS Program, Fiscal Years 2000-2007: [See PDF for image] This figure is a line graph depicting the following data: Fiscal year: 2000; Funding for the IRIS Program: $1.7 million. Fiscal year: 2001; Funding for the IRIS Program: $1.9 million. Fiscal year: 2002; Funding for the IRIS Program: $6.7 million. Fiscal year: 2003; Funding for the IRIS Program: $2.3 million. Fiscal year: 2004; Funding for the IRIS Program: $6.9 million. Fiscal year: 2005; Funding for the IRIS Program: $8.2 million. Fiscal year: 2006; Funding for the IRIS Program: $8.4 million. Fiscal year: 2007; Funding for the IRIS Program: $9.6 million. Source: GAO analysis of EPA data. Note: In fiscal year 2002, a congressional appropriations conference committee designated $5 million to accelerate the development of new IRIS values and to update current IRIS values. According to EPA officials, this funding was provided to various EPA program offices to support the IRIS assessments that program offices were leading at that time. In addition, EPA has reprogrammed funds from some of its other programs to expand the IRIS program to support the development of IRIS assessments, especially high-priority chemicals. [End of figure] Process Changes and Increased Resources Have Not Enabled EPA to Routinely Complete IRIS Assessments and Meet Users' Needs: As shown in figure 3, EPA has continued to take steps to improve the IRIS assessment process and increased program funding and staffing, but the agency has not made progress in completing IRIS assessments and reducing its backlog of ongoing assessments. Although the number of program staff has quadrupled from 8 to 37 between 2000 and 2007, EPA has, on average, completed about five IRIS assessments per year--and in fiscal years 2006 and 2007, completed only two each year. However, EPA sent 16 draft assessments to OMB for OMB/interagency review in both fiscal years 2006 and 2007 and plans to provide 16 draft assessments to OMB annually in fiscal years 2008 through 2012. Further, in its fiscal year 2008 budget justification documents, EPA said it plans to complete 8 IRIS assessments in fiscal year 2008, noting that completion of assessments, rather than providing drafts to OMB, is the most important outcome. Figure 3: Number of Completed IRIS Assessments, Draft Assessments to OMB, and IRIS Staff in Full-Time Equivalents, Fiscal Years 2000-2007: [See PDF for image] This figure is a multiple line graph depicting the following data: Fiscal year: 2000; IRIS staff levels in full-time equivalents: 7.8; IRIS assessments completed: 5. Fiscal year: 2001; IRIS staff levels in full-time equivalents: 7.8; IRIS assessments completed: 7. Fiscal year: 2002; IRIS staff levels in full-time equivalents: 7.8; IRIS assessments completed: 3. Fiscal year: 2003; IRIS staff levels in full-time equivalents: 12.8; IRIS assessments completed: 11. Fiscal year: 2004 (Year OMB/interagency review initiated); IRIS staff levels in full-time equivalents: 27; IRIS assessments completed: 4; Draft IRIS assessments to OMB for OMB/interagency review: 1. Fiscal year: 2005; IRIS staff levels in full-time equivalents: 37; IRIS assessments completed: 4; Draft IRIS assessments to OMB for OMB/interagency review: 5. Fiscal year: 2006; IRIS staff levels in full-time equivalents: 37; IRIS assessments completed: 2; Draft IRIS assessments to OMB for OMB/interagency review: 16. Fiscal year: 2007; IRIS staff levels in full-time equivalents: 37; IRIS assessments completed: 2. Draft IRIS assessments to OMB for OMB/interagency review: 16. Source: GAO analysis of EPA data. [End of figure] During fiscal year 2007, EPA had 77 ongoing IRIS assessments. As of December 1, 2007, 70 of these were still in progress--2 were completed in fiscal year 2007 and 5 were discontinued at the direction of OMB.[Footnote 6] Representing EPA's first efforts to develop IRIS assessments covering acute (short-term) exposure to chemicals, EPA initiated them in 2004 to meet EPA program needs, including, for example, to help it implement the Clean Air Act. In November 2007, EPA officials told the members of the Board of Scientific Counselors committee evaluating EPA's health risk assessment program that OMB did not explain to EPA why it was directing the agency to terminate these IRIS assessments. Of the 70 assessments still in progress as of December 1, 2007, 48 have been in progress for more than 5 years, and 12 of those for more than 9 years. As figure 4 shows, many of these assessments are either being drafted or are undergoing internal agency review. Additionally, 19 assessments are undergoing one of the two OMB/interagency reviews managed by OMB, and 5 of the assessments have been suspended by EPA for various reasons, including waiting for new research studies that have yet to be completed. Figure 4: December 2007 Status of IRIS Assessments That EPA Reported as Ongoing in Fiscal Year 2007: [See PDF for image] This figure is a vertical bar graph depicting the following data: Type: Suspended[A]; Number of assessments: 5. Type: Draft development[B]; Number of assessments: 31. Type: Formal EPA agency review[C]; Number of assessments: 6. Type: First interagency review managed by OMB[D]; Number of assessments: 14. Type: Assessments discontinued per OMB direction[E]; Number of assessments: 5. Type: Assessments delayed per EPA direction[F]; Number of assessments: 1. Type: External peer review[G]; Number of assessments: 8. Type: Second interagency review managed by OMB[H]; Number of assessments: 5. Type: Preparation of final assessment/web posting[I]; Number of assessments: 0. Type: Completed assessments[J]; Number of assessments: 2. Source: GAO analysis of EPA information. [A] EPA has suspended five IRIS assessments for several reasons, including waiting for additional research, reevaluating the need for an assessment, and potentially broadening the scope of an assessment. [B] This phase includes conducting literature searches, drafting assessments, and incorporating comments from an internal peer review process; it is concluded when the IRIS Program Director provides the draft and support documents for review by EPA's IRIS Agency Review Committee. [C] This phase includes the review by EPA's IRIS Agency Review Committee and revisions in response to the review; it is concluded when the NCEA Associate Director for Health transmits a draft to OMB. [D] This phase encompasses the first OMB/interagency review managed by OMB; revisions in response to the review; negotiations among EPA, OMB, and other federal agencies over changes to the draft and the content of the charge questions for the external peer reviewers; and briefing of the Assistant Administrator for the Office of Research and Development (ORD). It is concluded when OMB informs EPA that issues have been resolved and the assessment can proceed to the next step--external peer review. [E] In November 2007, EPA said that in 2006 OMB requested and then in 2007 directed EPA to discontinue the five IRIS assessments covering acute exposures that EPA had sent to OMB for OMB/interagency review from July to September 2006. [F] One assessment has been delayed by EPA for almost 2 years pending the development of an acceptable uncertainty analysis. [G] This phase encompasses public review and comment, the external peer review conducted by an independent panel, and incorporation of draft revisions in response to peer review and public comments. It is concluded when the NCEA Associate Director for Health transmits a revised draft to OMB for the second OMB/interagency review it manages. [H] This phase encompasses the second review by EPA's IRIS Agency Review Committee; the second OMB/interagency review managed by OMB; revisions in response to comments received from agency and OMB/ interagency review; negotiations among EPA, OMB, and the other federal agencies over changes; and briefing of the Assistant Administrator for ORD. It is concluded when ORD, the IRIS Agency Review Committee, and other federal agencies resolve all issues, OMB informs EPA that issues have been resolved, and a revised IRIS assessment is sent for preparation of a final IRIS assessment. [I] This phase encompasses the final preparation of the IRIS summary and toxicological review and the review and clearance by the NCEA Director. It is concluded when the assessment is posted on the IRIS Web site after the Assistant Administrator, ORD, approves a fact sheet about the assessment prepared by the chemical manager. [J] EPA completed two assessments in July and September 2007. [End of figure] In addition to the current backlog of 70 ongoing and suspended IRIS assessments, EPA data from 2001 through 2003 indicated that 287 of the chemicals in the IRIS database may potentially need to be updated. Specifically, EPA reviewed the scientific literature on the 460 chemicals in the database not being reassessed to identify assessments that may need to be updated in light of new studies or information that could potentially change the risk estimates currently in the IRIS assessments. In addition, while conducting these literature reviews, EPA identified new studies or information that would enable the agency to develop additional risk estimates (e.g., add an estimate, such as an RfD or cancer potency estimate, for assessments lacking such estimates). EPA's "screening level review" found new information that could potentially (1) change an existing risk estimate for 169 chemicals and/or (2) allow EPA to develop additional risk estimates for 210 chemicals. Although EPA identified these chemicals as candidates for reassessment, as of fiscal year 2007, the agency had initiated reassessments of only a few of these chemicals. Further, because the screening levels were completed more than 4 years ago, it is likely that more assessments may now be outdated. EPA officials said they have initiated another screening level review that they expect to be completed by July 2008. Importantly, in its performance assessment of EPA's human health risk assessment program, which includes IRIS, OMB considers health assessment values in IRIS as out of date if they were developed more than 10 years ago and where new scientific information has been identified that could change the health assessment value. This designation highlights the need for EPA to become productive in completing its IRIS assessments if it is to remain a viable and credible assessment resource. We note that EPA recently formed a new working group that will specifically address chemical assessments that are over 10 years old and for which new information is available. This group's efforts are currently in the planning stages, and an EPA official involved in this group indicated that the agency expects to begin work on this effort by March 2008. Although EPA has stated its intent that the IRIS database be updated and expanded to include new assessments requested by IRIS users as soon as "practically possible," the current backlog continues to present a practical impediment to doing so. Because of the backlog, for example, EPA did not initiate new assessments in fiscal years 2006 and 2007. [Footnote 7] In fact, key IRIS users--two EPA program offices-- specifically requested that the IRIS program management focus its resources on expediting the completion of ongoing IRIS assessments important to their regulatory and cleanup responsibilities rather than initiating new assessments in fiscal year 2007. For example, the Office of Solid Waste and Emergency Response identified 29 assessments that were important to that office and asked EPA to finish these IRIS assessments before initiating others. In addition, the Office of Air identified 28 high-priority IRIS assessments that it needed to fulfill its regulatory mandates,[Footnote 8] 14 of which it identified as being of the "highest priority" and the other as "high priority;" IRIS reviews for 12 of the chemicals of highest priority are in progress, but most of these assessments are still either being drafted or are undergoing an internal agency review. In terms of new assessments that will be needed for EPA's air toxics program, an official in the Office of Air and Radiation said there were about 12 to 20 air toxics of concern nationwide for which IRIS risk information is needed and about 50 or so uncommon air toxics emitted in isolated "hot spots" (e.g., near a chemical plant in sparsely populated areas) where a small number of people who live nearby--perhaps 200 or so--are exposed to relatively high doses of these air toxics. He said it is more difficult to get approval for assessments of such chemicals because of the small number of people potentially affected, even though these chemicals may be very harmful to human health and the environment. Moreover, as discussed above, EPA program offices and state and local entities have identified needs for assessments of hundreds of chemicals not yet in IRIS. For example, in its 2003 needs assessment report, EPA reported that about half of the chemicals and chemical classes that IRIS users had nominated for assessment were not included in IRIS. [Footnote 9] Reasons that IRIS users nominated these chemicals included to support anticipated rule making, to support agency or state implementation priorities, to address children's health concerns, and to address the widespread occurrence of a chemical at contaminated sites and in groundwater. Unmet IRIS needs affect IRIS data users both within and outside EPA. For example, EPA's Office of Research and Development noted in its 2003 Air Toxics Multi-Year Plan that quantitative assessment values (for noncancer and cancer health effects) for many high-priority air toxics that are needed to support site-specific efforts and regulatory decisions are "missing." Along these same lines, in its Office of Solid Waste Integrated Research and Development Plan for the Hazardous Waste Identification Rule, EPA's Office of Research and Development identified 460 chemicals of potential concern and reported that roughly 200 of these lack quantitative assessment values (estimates of the risk of cancer and noncancer effects of the chemicals). Non-EPA users also are affected by the lack of IRIS values. According to the preliminary results of an EPA project aimed at determining how IRIS is used by non-EPA decision makers,[Footnote 10] non-EPA users have one primary criticism regarding IRIS--they are frustrated by the lack of new assessments, particularly for "controversial" chemicals. These users reported that the absence of an IRIS assessment creates enormous challenges for state regulatory agencies and significant uncertainty for regulatory parties. Although EPA's 2003 needs assessment report had identified a potential need to complete 50 IRIS assessments annually to meet user needs, the agency has not finalized more than 11 assessments a year during the past 10 years. As figure 3 shows, productivity has, in fact, declined since 2003, with EPA completing 4 assessments in fiscal year 2005 and 2 each in fiscal years 2006 and 2007. EPA's updated multiyear plan estimates incremental increases in the number of IRIS assessments EPA will complete in fiscal years 2008 through 2011, at which point EPA's annual goal is to complete 16 IRIS assessments.[Footnote 11] We note that even if EPA were to overcome the significant productivity difficulties it has experienced in recent years and meet this goal of completing 16 assessments a year beginning in 2011, it is not clear that this level of productivity would meet the needs of EPA program offices and other users, given the current status of the database. For example, in November 2007, the Deputy Administrator of EPA's Region 2 told the Board of Scientific Counselors panel reviewing EPA's human health risk assessment program that even 16 IRIS assessments per year would not meet their chemical assessment needs. Although funding for the IRIS program has increased since 2000, we note that fiscal year 2007 funding of $9.6 million represents approximately 0.1 percent of EPA's $7.3 billion annual budget. It is a positive step that EPA delivered 32 IRIS assessments to OMB for OMB/interagency review in fiscal years 2006 and 2007. However, in general, the IRIS assessments sent to OMB during this period were on less controversial chemicals. In its December 2005 multiyear plan, EPA specified 10 major assessments that would be sent for external review in fiscal years 2006 and 2007. As of December 1, 2007, only 2 of these assessments had been sent for OMB/interagency review.[Footnote 12] In order to meet user needs and therefore enable EPA to more effectively protect public health, it will be important for EPA to make progress on the assessments of the more controversial chemicals, which tend to be those to which people are more widely exposed. Lastly, while EPA's assessment process changes have resulted in EPA providing the public with more information about IRIS assessments, some of the information provided in Federal Register notices and in the IRIS database itself is not accurate or is incomplete. For example, some assessments that have been suspended have continued to be identified as ongoing, and information on the status of the individual assessments provided in a system called IRIS Track has not been kept up to date. Regarding this latter point, we found that more EPA chemical managers started updating this database as a result of our review. Nonetheless, some of the milestone estimates remained outdated, and the milestones in IRIS Track did not reflect the current assessment process. We note that in November 2007, EPA created a new IRIS Web page, updating and consolidating some of the milestones to better reflect the current process. However, two key milestones--the second OMB/interagency and EPA internal agency reviews--are still not identified. EPA's Productivity Problems Stem from Several Key Factors: In our view, several key factors have contributed to EPA's inability to achieve a level of productivity that is needed to sustain the IRIS program and database. These factors that have hindered EPA's efforts to improve productivity are: * the OMB/interagency review process managed by OMB, * the growing complexity and scope of risk assessments, * certain management decisions and issues regarding the IRIS program, * congressional action that has delayed some assessments with potentially significant economic effects, and: * the compounding effect of delays. OMB/Interagency Review Process: One factor that has made it more difficult for EPA to complete IRIS assessments in a timely manner is the OMB-managed interagency review process, initiated in 2004 at OMB's direction. According to OMB, the purpose of OMB/interagency reviews is to ensure that federal agencies are aware of draft IRIS assessments in which they have an interest and that these agencies have the opportunity to be involved with the IRIS assessments as they are being developed in order to help ensure and increase their quality. This process, initially conducted on an ad hoc basis, was put in place in response to interagency conflicts that EPA faced when it attempted to finalize some IRIS assessments for chemicals that became highly controversial, such as perchlorate, naphthalene, and TCE--chemicals that are or have been considered by some federal agencies, including DOD, DOE, and NASA, to be integral to their missions. Notably, EPA's IRIS assessments of these chemicals could lead to regulatory actions that could, among other things, restrict the use of these chemicals, require agencies to provide protective gear to their employees exposed to the chemicals at work, or require agencies or their contractors to carry out or pay for cleanup of contamination at federal sites. The interagency conflicts about these IRIS assessments have contributed to their delays--resulting, for example, in EPA having to essentially restart the naphthalene assessment after it had been drafted and peer reviewed. The OMB/interagency review process has evolved over time, but it remains an informal process with OMB generally communicating its review requirements verbally. OMB has not provided EPA with written guidance, directives, policies, or procedures on this significant review program. In 2005, OMB started requiring EPA to send OMB all draft IRIS assessments for an OMB/interagency review before the assessments are provided to external peer review panels. In addition, OMB requires EPA to send the assessments to OMB a second time, after EPA has addressed the external peer review comments and recommendations. According to EPA officials, in 2007, OMB informed EPA that it cannot send its draft assessments for external peer review or post final EPA assessments on the EPA IRIS database until OMB verbally informs EPA that it has satisfactorily addressed OMB/interagency comments. OMB has not specified which authority or authorities it is using to review IRIS assessments. Because IRIS assessments are not regulations, they are not covered by Executive Order 12866 which provides for OMB review of proposed regulations, among other things. In addition, although in January 2007, Executive Order 13422 amended Executive Order 12866 by establishing a role for OMB in reviewing "significant guidance" documents, which could potentially include IRIS assessments, OMB officials told us that OMB has not formally classified IRIS assessments as significant guidance documents within the meaning of this executive order.[Footnote 13] These officials said that OMB had the authority to review IRIS assessments prior to the issuance of Executive Order 13422 and was continuing to use this general authority. We note that these executive orders address only OMB's reviews of two specific categories of agency documents and that OMB has not identified IRIS assessments as falling into either of these categories. Under the OMB/interagency review process for IRIS assessments, OMB identifies the federal agencies from which it will seek comments on the assessments and provides comments and questions to EPA from OMB and the other federal agencies. EPA then revises the assessments to address the comments and questions and also provides OMB with a document discussing how it has addressed each of the issues raised. According to EPA officials, this step is concluded when OMB verbally informs EPA that all of the issues are resolved. Thus, before EPA can provide its draft assessments to external peer reviewers, EPA, OMB, and the other federal agencies must reach agreement on (1) EPA's revised IRIS assessment and (2) the scope of the external peer review, including specific questions reviewers will be asked. The OMB reviews have primarily been conducted by a policy analyst in the Office of Information and Regulatory Affairs who, according to OMB officials, has a toxicology background and is qualified to conduct these reviews. If EPA does not agree with some of the proposed changes or peer review requirements, OMB officials said the assessment could be elevated to higher management levels within EPA for resolution. Thus, the framework of the OMB/interagency review process can essentially give more weight to OMB and the other federal agencies than to EPA chemical managers who have prepared the assessments and the many scientists and experts who have peer reviewed them. Further, in commenting on OMB's proposed risk assessment bulletin in 2007, the National Academies stated its concern that scientific issues may be superseded by policy considerations to the extent that the technical aspects of risk assessments would be overseen by OMB and not by the peer review process or by agency technical managers. Similarly, an EPA official noted that the farther removed the scientists and experts who have prepared or peer reviewed the assessments are from the negotiations and decisions over assessment changes requested by OMB and other federal agencies, the decisions are based more on political rather than scientific considerations. According to EPA officials, some of the OMB/interagency reviews have provided valuable input. However, they said the reviews have also added a significant amount of time to the assessment process, in part because the reviews are not subject to any specific time frames or deadlines and because responding to OMB comments in some cases has required several iterations to address nonsubstantive issues involving minor clarifications, semantics, and organization. (We note that according to the executive order under which OMB reviews regulatory actions, OMB must generally complete such reviews within 90 days.) In terms of time frames, our analysis of EPA data as of December 1, 2007, indicate that of the 36 assessments sent to OMB for the first OMB/interagency review, 16 have completed this process. While 4 completed the process in less than 6 months, 12 of them were in the review process for 6 or more months. Moreover, five additional assessments--the previously discussed assessments of acute exposure--were discontinued at the direction of OMB after a year in the OMB/interagency review stage. In addition, as of December 1, 2007, 5 of 10 draft assessments had completed the second OMB/interagency review, which OMB officials said is conducted by OMB to ensure that EPA has adequately considered the comments from external peer review panels. These assessments were at the second OMB/interagency review stage for periods ranging from 10 days to almost 4 months. According to EPA, while agencies should only point out major scientific issues that would warrant halting the release of the assessment, some comments submitted to them from the second OMB/interagency reviews by OMB have gone beyond their intended purpose of identifying only major scientific concerns and thus have unnecessarily added time to the assessment process. (See app. III for additional information about the time frames for the OMB/interagency reviews.) Further, EPA officials noted that the OMB/interagency review process can delay not only the IRIS assessments undergoing OMB/ interagency review but also the other assessments that the EPA chemical managers would be working on or starting if they were not engaged in responding to the OMB/interagency comments. Overall, the two OMB/interagency reviews managed by OMB have introduced a significant level of uncertainty into the time frames for completing IRIS assessments. This fact was reflected in EPA's December 2005 Human Health Risk Assessment Multi-Year Plan, which identified EPA's annual performance goals for fiscal years 2006 through 2012. The goals did not include the number of IRIS assessments completed and posted on the IRIS Web page--the performance goal that would be expected--but instead the number of assessments provided to OMB for the first OMB/interagency review. While EPA's annual performance goal for IRIS assessments had been the number of completed assessments, during the Program Assessment Rating Tool (PART) review by OMB, it was agreed that the number of assessments provided to OMB for OMB/interagency review was the most appropriate annual measure of performance because EPA "relinquishes direct control of production dates" when it sends draft IRIS assessments to OMB. In addition to adding time to the IRIS assessment process, the OMB/ interagency review process also affects the credibility of assessments primarily because the review process lacks transparency. According to EPA's Risk Characterization Policy and Handbook, the risk assessment process--which includes the hazard identification and dose-response analysis that constitutes IRIS assessments--should be transparent and its products should be clear, consistent, and reasonable. Transparency is particularly important in cases such as this when potentially affected parties are providing input into, and in some cases questioning, EPA's scientific analyses supporting its IRIS assessments. Specifically, a transparent process allowing IRIS users and the public to see the comments from OMB and other federal agencies--including those potentially impacted by the IRIS assessments--as well as EPA's responses to them, could help alleviate concerns about potential bias in the assessments. However, under the OMB/interagency review process managed by OMB, the comments EPA receives from OMB and other federal agencies and EPA's responses to them are not available to the public. OMB does not authorize their disclosure to the public on the basis that these communications are internal deliberations of the executive branch. Overall, because the rationale for changes to EPA's scientific assessments stemming from OMB and the interagency review process are not disclosed, the credibility of the IRIS assessments is reduced. We note that the former Assistant Administrator for the Office of Research and Development has emphasized the importance of transparency in the IRIS assessment process. Specifically, he stated that the best cure for controversy surrounding IRIS assessments "is early and frequent visits to the experts, second opinions, and lots of sunshine" (that is, transparency). Growing Complexity of Risk Assessments and Risk Assessment Methods and Models: Another factor that affects the length of time it takes to complete IRIS assessments is the growing scientific complexity of the assessments. For example, according to John Graham, the former director of both Harvard's Center for Risk Analysis and OMB's Office of Information and Regulatory Affairs, more and different types of scientific data have made the IRIS assessment of health effects more challenging.[Footnote 14] As a result, EPA chemical managers and scientists need to explore new methods of analyses and evaluate a wider variety of potential health effects (e.g., multiple disease endpoints) than in the past. In addition, chemical managers responsible for the assessments are working with a growing body of complex risk assessment guidelines, such as EPA's 2005 Guidelines for Carcinogen Risk Assessment (final revised cancer assessment guidelines) and its supplement relating to children, Supplemental Guidance for Assessing Susceptibility from Early Life to Exposure to Carcinogens. In addition, chemical managers must increasingly analyze studies using state-of-the- art physiologically based pharmacokinetic (PBPK) models, which improve the estimation of doses across species and routes of exposure and provide insights on uncertainty.[Footnote 15] Before EPA relies on PBPK models intended for risk assessments, the agency needs to evaluate them. Such evaluations include a review of the model purpose, structure, mathematical representation, parameter estimation (calibration), and computer implementation. EPA has established criteria for acceptance of a PBPK model for risk assessment purposes in one of two detailed reports issued in 2006, Approaches for the Application of Physiologically Based Pharmacokinetic (PBPK) Models and Supporting Data in Risk Assessment. The other 2006 report on PBPK models is the Use of Physiologically Based Pharmacokinetic (PBPK) Models to Quantify the Impact of Human Age and Interindividual Differences in Physiology and Biochemistry Pertinent to Risk. In fiscal years 2008 and 2010, EPA plans to issue additional information on PBPK modeling for use in risk assessments. The chemical managers are also called upon to use some new methods and models of risk assessment that are being developed or being applied for the first time in the absence of guidelines. For example, uncertainty analysis is a method in the early stages of development and use in risks assessments, for which EPA has not yet developed guidance. [Footnote 16] However, chemical managers are having to develop and apply new approaches to quantify and communicate uncertainty. According to EPA, these approaches include identifying alternative studies and endpoints for the application of uncertainty factors in noncancer risk assessment and the application of alternative dose- response models for cancer risk assessment. Important cutting-edge assessment models relevant to IRIS assessments include biologically based dose-response models, which are based on the modes of action of chemicals--that is, analysis of physiological, chemical, and biological information that helps identify a chemical agent's role in the development of tumors. These highly complex models can analyze multiple modes of action and can be responsive to EPA's 2005 guidelines for cancer assessment that emphasize mode-of-action evaluation. According to EPA, the agency has extensive experience in qualitative mode-of- action analysis but more limited experience in the quantitative evaluation of multiple modes of action. EPA Management Decisions and Issues: Some of the IRIS assessment process changes that EPA management has implemented in recent years have made it more difficult for the agency to complete assessments in a timely manner. These changes were aimed primarily at improving the quality of the assessments. It is too early to determine whether or to what extent the changes have enhanced the scientific credibility of the assessments,[Footnote 17] but it is clear they have contributed to assessment delays overall. Changes that have affected the time frames for IRIS assessments include waiting for additional scientific studies to be completed, waiting for the development of an acceptable methodology for presenting qualitative and quantitative uncertainty analysis in the IRIS assessments, and numerous process changes. Waiting for additional scientific studies: EPA management's decision in some cases to suspend ongoing IRIS assessments while waiting for additional scientific studies to be completed has contributed to EPA's lack of productivity. According to a former IRIS program director, EPA's general approach in the 1990s was to use only information from completed scientific studies available at the time of the assessment-- e.g., the reviews were based on the best available science. However, EPA has awaited the results of new and ongoing studies before completing some IRIS assessments, which has resulted in delaying them for years. Examples of key chemical assessments that have been delayed while EPA waits for new studies include those for formaldehyde and Royal Demolition Explosive,[Footnote 18] discussed later. Other delayed assessments include those for tetrahydrofuran, perfluorooctane sulfonate-potassium salt (PFOS), and perfluorooctanoic acid-ammonium salt (PFOA). We understand that there may be exceptional circumstances for which it may be appropriate to wait for the results of an important ongoing study, such as a major epidemiological study that will provide new, critical data for an assessment. According to EPA officials, this is the case with research they are awaiting for its IRIS assessment of asbestos. However, as a general rule, requiring that IRIS assessments be based on the best science available at the time of the assessment, as had been the prior practice, is a standard that would best support a goal of completing assessments within reasonable time periods and minimizing the need to conduct significant levels of rework. Developing an acceptable uncertainty analysis for IRIS assessments: Another management decision that has delayed the completion of IRIS assessments is to incorporate comprehensive uncertainty analysis into the IRIS assessments. Peer reviews of EPA's assessments by the National Academies and others have sometimes recommended additional uncertainty analysis; the Assistant Administrator, Office of Research and Development, has made the inclusion of qualitative and quantitative uncertainty analyses in IRIS assessments a high priority to, among other things, support better decisions and guide EPA's research agenda. Such analyses require the use of state-of-the-art tools to quantify uncertainty. That is, comprehensive--and, in particular, quantified-- uncertainty analysis, is an emerging analytic tool. For example, EPA's August 2007 Human Health Risk Assessment Multi-Year Plan estimates releasing an external review draft report on methods for analyzing and characterizing uncertainty in hazard and dose-response assessments in 2010; the plan--which covers 2006 through 2012--does not estimate a final report date for this important guidance on uncertainty analysis. In the interim, EPA chemical managers have had to try to develop complex uncertainty estimates in draft assessments in the absence of agency guidance or protocols. As discussed later, this requirement has delayed the completion of an important assessment (tetrachloroethylene) for almost two years. Moreover, because EPA is now requiring all IRIS assessments to include basic or enhanced uncertainty analysis, other significant assessments that have been drafted have also been delayed, pending approval by the Assistant Administrator, Office of Research and Development, of an acceptable template for uncertainty analysis for significant assessments. Further, external peer reviewers of IRIS assessments containing quantified uncertainty analyses will need to have specialized expertise to assess the quality and reliability of these cutting-edge analyses--which themselves contain uncertainties and incorporate numerous assumptions. Evaluating the uncertainty assessments will be challenging, given their complexities and the lack of guidance on this emerging method. Along these lines, members of the Board of Scientific Counselors[Footnote 19] have questioned whether expert peer reviewers of EPA's IRIS assessments will have the necessary expertise in this emerging method. Finally, effectively communicating the results of complex uncertainty analyses may be challenging--in fact, some have questioned whether highly detailed quantitative uncertainty analysis enhances the values of risk assessments. For example, in commenting on a proposed OMB bulletin that would have provided risk assessment guidance to federal agencies, the Department of Health and Human Services noted that "characterization of every possible uncertainty or extensive evaluation of each assumption . . . could result in a confusing, less straight-forward document . . . that would not serve the public or the risk assessment community well." [Footnote 20] Similarly, in its review of the proposed bulletin, the National Academies concluded that "there is a serious danger that agencies will produce ranges of meaningless and confusing risk estimates, which could result in risk assessments of reduced rather than enhanced quality and objectivity." Continuous process changes, outdated standard operating procedures, and management issues: EPA's continual changes to the IRIS assessment process present a challenge to the chemical managers who are undertaking the assessments. Further, a number of changes have been implemented informally since the last update to the agency's standard operating procedures for fiscal year 2006 reviews. According to EPA, these changes have been made in order to continually improve the assessment process and to respond to changing requirements, such as the OMB-managed OMB/interagency review and the need to incorporate emerging risk assessment methods. Chemical managers have told us that there is much confusion about the process, including who needs to be briefed on the draft assessments and who has approval authority to move an assessment to the next step. In fact, in the absence of current operating procedures to guide chemical managers on basic procedures and program management responsibilities for the development, review, and finalization of IRIS assessments, EPA chemical managers recently took the initiative to develop a detailed graphic providing their best understanding of the process. One aspect of the IRIS assessment process that has become particularly unclear is the steps required to obtain approvals to move draft assessments through various stages (e.g., to OMB, external peer review, or final issuance). For example, until 2004, EPA used an agencywide consensus process to reach either unanimous or substantial agreement on draft assessments among members of the IRIS Agency Review Committee; if agreement could not be reached by the committee, the operating procedures had provided a detailed process for resolving contentious issues. When EPA moved from the consensus approach to an "agency resolution" approach, however, the guidance said only that final decisions on any contentious issues were to be resolved by either the NCEA Director or the Agency Science Advisor.[Footnote 21] In addition, while approval to move draft assessments to external peer review had been made by the NCEA Director, according to EPA officials, in 2006, the Assistant Administrator started approving drafts for the first of three external reviews of IRIS assessments--the initial OMB/ interagency review. However, this change occurred informally and was not reflected in the existing standard operating procedures.[Footnote 22] We also note that, starting in 2004, chemical managers were required to provide a fact sheet on final IRIS assessments to the Assistant Administrator for review and approval before the assessments could be posted on the Web site.[Footnote 23] Raising the level at which various approvals are made has added time to the process because it involves additional briefings, additional revisions and negotiations over revisions, and, at times, delays due to scheduling difficulties. As a result, it is important for EPA to both articulate and evaluate its internal review process to help ensure the appropriate balance between product quality and timeliness. Finally, in addition to process and approval clarifications, it also appears that the IRIS program could benefit from more rigorous management attention and oversight. For example, the risk assessment for chloroform--identified as one of the highest-priority assessments requested by the EPA Office of Air and Radiation--was first announced in a January 1998 Federal Register notice; however, the chemical manager for the chloroform assessment retired from the agency in late 2004, and a new chemical manager was not assigned to this assessment until early 2007.[Footnote 24] Although EPA continued to identify this assessment as ongoing, it was actually unstaffed and therefore suspended for at least 2 years. In addition, a relatively simple, less controversial assessment of 2,2,4-trimethylpentane, which concluded that there were insufficient scientific data to either quantify the noncancer health risks associated with the chemical or to assess its potential to cause cancer, took more than 4 years to complete. EPA officials could not explain why this assessment--which should have taken less than a year to complete--took so long, aside from noting that the responsible staff had been given other high-priority assignments during this time. EPA needs to ensure that its IRIS assessments are given high priority and adequately staffed so that costly delays are avoided. Congressional Actions: Another factor that can delay the completion of IRIS assessments is the potential for congressional involvement in assessments that become controversial, such as those with potentially significant economic effects. Because of the potential for such assessments to lead to regulatory actions that can significantly affect certain industries or federal agencies, it is particularly important that these assessments effectively and appropriately use the best available science. Even assuming the best available science, however, uncertainties remain an inherent aspect of complex chemical risk assessments. As the National Academies has noted, "A risk assessment usually involves incomplete data, scientific uncertainty, and the need for expert judgment" and "almost every risk assessment is open to challenge on one ground or another."[Footnote 25] As a result, the assessments may be questioned on various scientific and technical grounds and subjected to intense national scrutiny by many individuals and organizations, including the media. Constituents may contact their elected representatives with their concerns, and politicians are likely to become involved, either supporting the content of the assessments or challenging it. Because of this debate, Members of the Congress, congressional committees, or the Congress as a whole may direct EPA to take certain steps before finalizing a particular assessment. While the intense scrutiny to which some chemical assessments are subjected can result in improved assessments, all uncertainties cannot be eliminated and controversies can continue beyond the point at which additional analysis is helpful. Further, addressing questions and concerns about assessments often involves considerable rework and takes a significant amount of time. EPA strives to balance the desire for the best possible assessment with its responsibilities for protecting the public health, which it can only do with timely assessments of chemicals. That being said, it can be challenging for EPA to both identify and achieve the proper balance between these competing goals. In the case of certain controversial chemical assessments, actions by congressional committees and individual members have led EPA to, for example, postpone completion of the IRIS assessment of formaldehyde for years until an update of an epidemiological study that had just been released was completed. Another response to congressional concerns is EPA's decision to reconsider the quantitative noncancer assessment of a chemical, dibutyl phthalate, that had completed all internal and external reviews and was ready to be released in 2007. The noncancer assessment, an update to the assessment completed in 1990, would have allowed more exposure to dibutyl phthalate than the earlier assessment. In 2000, the Department of Health and Human Services' National Toxicology Program concluded that dibutyl phthalate may adversely affect human reproduction or development if exposures are sufficiently high. This chemical, regulated under the Clean Air Act's air toxics program, is used in many consumer products, such as cosmetics, nail polishes, fragrances, wood stains, and toys; the European Union has banned its use in cosmetics and restricted its use in children's toys containing concentrations of phthalates in excess of 0.1 percent. Similarly, after January 1, 2009, a California statute enacted in 2007 will prohibit the manufacture, sale, or distribution of certain toys and child care articles if the products contain concentrations of phthalates exceeding 0.1 percent.[Footnote 26] Because of congressional questions and comments from the Natural Resources Defense Council about, among other things, the adequacy of the uncertainty factors used in the analysis and the lack of consideration in the draft assessment of cumulative exposure to multiple phthalates (similar chemicals in the same class) that have similar health effects, EPA has suspended the assessment. In December 2007, an EPA official said that EPA has decided to seek advice from the National Academies on whether the agency should develop a new type of assessment for this chemical--one that would assess a class of related chemicals. If EPA decides to develop a new, cumulative IRIS assessment, it will likely be many years before an updated IRIS assessment addressing dibutyl phthalate is completed. [Footnote 27] Compounding Effect of Delays: An overarching factor that affects EPA's ability to complete IRIS assessments in a timely manner is that once a delay in the assessment process occurs, work that has been completed can become outdated, necessitating rework throughout some or all of the assessment process. For example, delays often require repeating reviews of the scientific literature on a chemical to take into account the time that has passed since the literature review was completed; this, in turn, may require detailed analyses of any new studies found to be relevant. Moreover, new risk assessment guidelines and cutting-edge methodologies that the agency has started applying to other assessments may now need to be applied to an assessment being reworked to meet current assessment requirements and standards and the expectations of peer reviewers. Once these analyses are complete, the draft will likely need to be revised and again subjected to internal reviews, OMB/interagency reviews, and scientific peer review. In addition, the longer an assessment is in progress, the more likely it becomes that staff will change due to retirements and resignations; in turn, newly assigned chemical managers face a learning curve in reviewing often voluminous data and analyses. For example, at least four chemical managers have been assigned to the nitrobenzene assessment since it was started in 1998. Overall, even a single delay can have a domino effect--having far-reaching, time- consuming consequences, in some cases, requiring that the assessment process essentially start over. In addition, because chemical managers must continue to devote time and attention to assessments that experience delays--often repeating prior steps to update the assessments--their ability to work on other ongoing assessments and undertake new ones is limited. Key IRIS Assessments Have Been Delayed by Some of These Factors: Some key IRIS assessments have been in progress for a number of years, in part because of delays stemming from one or more of the factors discussed above. Examples include the following: Naphthalene. EPA started the IRIS assessment of cancer risks stemming from the inhalation of naphthalene in 2002. Naphthalene is used in jet fuel and in the production of widely used commercial products such as moth balls, dyes, insecticides, and plasticizers. According to a presentation delivered at the 2007 annual meeting of the Society for Risk Analysis by an Army Corps of Engineers toxicologist,[Footnote 28] "The changing naphthalene regulatory environment includes a draft EPA risk assessment that if/when finalized, will change naphthalene's status from 'possible' to 'likely' human carcinogen."[Footnote 29] Thus, according to this presentation, one potential impact of this IRIS assessment on DOD is that DOD would need to provide many employees exposed to naphthalene with equipment measuring their exposure to the chemical. In addition, because many military bases are contaminated with naphthalene, a component of jet fuel (approximately 1 percent to 3 percent) used by all DOD services, DOD could face extensive cleanup costs. By 2004, 2 years after starting the assessment, EPA had drafted a chemical assessment that had completed internal peer reviews and was about to be sent to an external peer review committee. Once it returned from external review, the next step, at that time, would have been a formal review by EPA's IRIS Agency Review Committee. If approved, the assessment would have been completed and released. However, in part because of concerns raised by DOD, OMB asked to review the assessment and conducted an interagency review of the draft. In their 2004 reviews of the draft IRIS assessment, both OMB and DOD raised a number of concerns about the assessment and suggested to EPA that it be suspended until additional research could be completed to address what they considered to be significant uncertainties associated with the assessment. Although all of the issues raised by OMB and DOD were not resolved, EPA continued with its assessment by submitting the draft for external peer review, which was completed in September 2004.[Footnote 30] However, according to EPA, OMB continued to object to the draft IRIS assessment and directed EPA to convene an additional expert review panel on genotoxicity to obtain recommendations about short-term tests that OMB thought could be done quickly.[Footnote 31] According to EPA, this added 6 months to the process, and the panel, which met in April 2005, concluded that the research that OMB was proposing could not be conducted in the short term. Nonetheless, EPA officials said that the second expert panel review did not eliminate OMB's concerns regarding the assessment, which they described as reaching a stalemate. In September 2006, EPA decided, however, to proceed with developing the assessment. By this time, the naphthalene assessment had been in progress for over 4 years; EPA decided that the IRIS noncancer assessment, issued in 1998, was outdated and needed to be revisited. Thus, EPA expanded the IRIS naphthalene assessment to include both noncancer and cancer assessments. As a result, 6 years after the naphthalene assessment began, it is now back at the drafting stage. The assessment now will need to reflect relevant research completed since the draft underwent initial external peer review in 2004, and it will have to undergo all of the IRIS assessment steps again, including additional internal and external reviews that are now required. This series of delays has limited EPA's ability to conduct its mission. For example, the Office of Air and Radiation has identified the naphthalene assessment as one of its highest-priority needs for its air toxics program. In addition, the Office of Solid Waste and Emergency Response considers the naphthalene assessment a high priority for the Superfund program--naphthalene has been found in at least 654 of Superfund's current or former National Priorities List sites.[Footnote 32] Although EPA currently estimates that it will complete the assessment in 2009, meeting this revised estimate will be challenging, given all of the steps that are yet to be completed and the extensive external scrutiny to which it will continue to be subjected. Royal Demolition Explosive. This chemical, also called RDX or hexahydro- 1,3,5-trinitrotriazine, is a highly powerful explosive used by the U.S. military in thousands of munitions. Currently classified by EPA as a possible human carcinogen, this chemical is known to leach from soil to groundwater. Royal Demolition Explosive can cause seizures in humans and animals when large amounts are inhaled or ingested, but the effects of long-term, low-level exposure on the nervous system are unknown. As is the case with naphthalene, the IRIS assessment could potentially require DOD to undertake a number of actions, including steps to protect its employees from the effects of this chemical and to clean up many contaminated sites. Although EPA started an IRIS assessment of Royal Demolition Explosive in 2000, it has made minimal progress on the assessment because EPA agreed to a request by DOD to wait for the results of DOD-sponsored research on this chemical. In 2007, EPA began to actively work on this assessment, although some of the DOD-sponsored research is still outstanding. Formaldehyde. EPA began an IRIS assessment of formaldehyde in 1997 because the existing assessment was determined to be outdated.[Footnote 33] Formaldehyde is a colorless, flammable, strong-smelling gas used to manufacture building materials, such as pressed wood products, and used in many household products, including paper, pharmaceuticals, and leather goods. While EPA currently classifies formaldehyde as a probable human carcinogen, the International Agency for Research on Cancer (IARC), part of the World Health Organization, classifies formaldehyde as a known human carcinogen. Since 1986, studies of industrial of workers have suggested that formaldehyde exposure is associated with nasopharyngeal cancer, and possibly with leukemia. For example, in 2003 and 2004, the National Cancer Institute (NCI) and the National Institute of Occupational Safety and Health (NIOSH) released epidemiological studies following up on earlier studies tracking about 26,000 and 11,000 industrial workers, respectively, exposed to formaldehyde; the updates showed exposure to formaldehyde might also cause leukemia in humans, in addition to the cancer types previously identified. According to NCI officials, the key findings in their follow-up study were an increase in leukemia deaths and, more significantly, an exposure/response relationship between formaldehyde and leukemia--as exposure increased, the incidence of leukemia also rose. As with the earlier study, NCI found more cases of a rare form of cancer, nasopharyngeal cancer, than would usually be expected. The studies from NCI and NIOSH were published in 2003 and 2004,[Footnote 34] around the time that EPA was still drafting its IRIS assessment. In November 2004, the Chairman of the Senate Environment and Public Works Committee requested that EPA delay completion of its IRIS assessment until an update to the just-released NCI study could be conducted, indicating that the effort would take, at most, 18 months. EPA agreed to wait--and more than 3 years later, the NCI update is not yet complete. As of December 2007, NCI estimates that the study will be completed in two stages, one in mid-2008 and the second one later that year. An NCI official said that the additional leukemia deaths identified in the update provide "greater power" to detect associations between exposure to formaldehyde and cancer. EPA's inability to complete the IRIS assessment it started more than 10 years ago in a timely manner has had a significant impact on EPA's air toxics program. Specifically, when EPA promulgated a national emissions standard for hazardous air pollutants covering facilities in the plywood and composite wood industries in 2004, EPA's Office of Air and Radiation took the unusual step of not using the existing IRIS estimate but rather decided to use a cancer risk estimate developed by an industry- funded organization, the CIIT Centers for Health Research (formerly, the Chemical Industry Institute of Toxicology) that had been used by the Canadian health protection agency. The IRIS cancer risk factor had been subject to criticism because it was last revised in 1991 and was based on data from the 1980s. In its final rule, EPA stated that "the dose-response value in IRIS is based on a 1987 study, and no longer represents the best available science in the peer-reviewed literature." The CIIT quantitative cancer risk estimate that EPA used in its health risk assessment in the plywood and composite wood national emissions standard indicates a potency about 2,400 times lower than the estimate in IRIS that was being re-evaluated and that did not yet consider the 2003 and 2004 NCI and NIOSH epidemiological studies. According to an EPA official, an IRIS cancer risk factor based on the 2003 and 2004 NCI and NIOSH studies would likely be close to the current IRIS assessment, which EPA has been attempting to update since 1997. The decision to use the CIIT assessment in the plywood national emissions standard was controversial, and officials in EPA's National Center for Environmental Assessment said the center identified numerous problems with the CIIT estimate. Nonetheless, the Office of Air and Radiation used the CIIT value, and that decision was a factor in EPA exempting certain facilities with formaldehyde emissions from the national emissions standard. In June 2007, a federal appellate court struck down the rule, holding that EPA's decision to exempt certain facilities that EPA asserted presented a low health risk exceeded the agency's authority under the Clean Air Act.[Footnote 35] Further, the continued delays of the IRIS assessment of formaldehyde--currently estimated to be completed in 2010 but after almost 11 years still in the draft development stage--will impact the quality of other EPA regulatory actions, including other air toxics rules and requirements. Trichloroethylene. Also known as TCE, this chemical is a solvent widely used as a degreasing agent in industrial and manufacturing settings; it is a common environmental contaminant in air, soil, surface water, and groundwater. TCE has been linked to cancer, including childhood cancer, and other significant health hazards, such as birth defects. TCE is the most frequently reported organic contaminant in groundwater, and contaminated drinking water has been found at Camp Lejeune, a large Marine Corps base in North Carolina. TCE has also been found at Superfund sites and at many industrial and government facilities, including aircraft and spacecraft manufacturing operations. In 1995, the International Agency for Research on Cancer classified TCE as a probable human carcinogen, and in 2000, the Department of Health and Human Services' National Toxicology Program concluded that it is reasonably anticipated to be a human carcinogen. Because of questions raised by peer reviewers about the IRIS cancer assessment for TCE, EPA withdrew it from IRIS in 1989 but did not initiate a new TCE cancer assessment until 1998. In 2001, EPA issued a draft IRIS assessment for TCE that proposed a range of toxicity values indicating a higher potency than in the prior IRIS values and characterizing TCE as "highly likely to produce cancer in humans." The draft assessment, which became controversial, was peer reviewed by EPA's Scientific Advisory Board and released for public comment. A number of scientific issues were raised during the course of these reviews, including how EPA had applied emerging risk assessment methods--such as assessing cumulative effects (of TCE and its metabolites) and using a physiologically based pharmacokinetic model--and the uncertainty associated with the new methods themselves.[Footnote 36] To help address these issues, EPA, DOD, DOE, and NASA sponsored a National Academies review to provide guidance. The National Academies report, which was issued in 2006, concluded that the weight of evidence of cancer and other health risks from TCE exposure had strengthened since 2001 and recommended that the risk assessment be finalized with currently available data so that risk management decisions could be made expeditiously. The report specifically noted that while some additional information would allow for more precise estimates of risk, this information was not necessary for developing a credible risk assessment. Nonetheless, 10 years after EPA started its IRIS assessment, the TCE assessment is back at the draft development stage. EPA estimates this assessment will be finalized in 2010. More in line with the National Academies' recommendation to act expeditiously, five senators introduced a bill in August 2007 that, among other things, would require EPA to both establish IRIS values for TCE and issue final drinking water standards for this contaminant within 18 months. Tetrachloroethylene. EPA started an IRIS assessment of tetrachloroethylene--also called perchloroethylene or "perc"--in 1998. Tetrachloroethylene is a manufactured chemical widely used for dry cleaning of fabrics, metal degreasing, and making some consumer products and other chemicals. Tetrachloroethylene is a widespread groundwater contaminant, and the Department of Health and Human Services' National Toxicology Program has determined that it is reasonably anticipated to be a carcinogen. The IRIS database currently contains a 1988 noncancer assessment based on oral exposure that will be updated in the ongoing assessment. Importantly, the ongoing assessment will also provide a noncancer inhalation risk and a cancer assessment. The IRIS agency review of the draft assessment was completed in February 2005, the draft assessment was sent to OMB for OMB/interagency review in September 2005, and the OMB/interagency review was completed in March 2006. EPA had determined to have the next step, external peer review, conducted by the National Academies--the peer review choice reserved for chemical assessments that are particularly significant or controversial. EPA contracted with the National Academies for a review by an expert panel, and the review was scheduled to start in June 2006 and be completed in 15 months. However, as of December 2007, the draft assessment has not yet been provided to the National Academies. After verbally agreeing with both the noncancer and cancer assessments following briefings on the assessments, the Assistant Administrator, Office of Research and Development, subsequently requested that additional uncertainty analyses--including some quantitative analyses--be conducted and included in the assessment before the draft was released to the National Academies for peer review. As discussed above, quantitative uncertainty analysis is a risk assessment tool that is currently being developed, and although the agency is working on developing policies and procedures for uncertainty analysis, such guidance currently does not exist. The draft tetrachloroethylene assessment has been delayed since early 2006 as EPA staff have gone back and forth with the Assistant Administrator trying to reach agreement on key issues such as whether a linear or nonlinear model is most appropriate for the cancer assessment and how uncertainty should be qualitatively and quantitatively characterized. EPA officials and staff noted that some of the most experienced staff are being used for these efforts, limiting their ability to work on other IRIS assessments. In addition, the significant delay has impacted the planned National Academies peer review because the current contract, which has already been extended once, cannot be extended beyond December 2008. The peer review was initially estimated to take 15 months. As a result, a new contract and the appointment of another panel may be required. Dioxin. The dioxin assessment is an example of an IRIS assessment that has been, and will likely continue to be, a political as well as a scientific issue. Often the byproducts of combustion and other industrial processes, complex mixtures of dioxins enter the food chain and human diet through emissions into the air that settle on soil, plants, and water. EPA's initial dioxin assessment, published in 1985, focused on the dioxin TCDD (2,3,7,8-tetrachlorodibenzo-p-dioxin) because animal studies in the 1970s showed it to be the most potent cancer-causing chemical studied to date. Several years later, EPA decided to conduct a reassessment of dioxin because of major advances that had occurred in the scientific understanding of dioxin toxicity and significant new studies on dioxins' potential adverse health effects. Initially started in 1991, this assessment has involved repeated literature searches and peer reviews. For example, a draft of the updated assessment was reviewed by a scientific peer review panel in 1995, and three panels reviewed key segments of later versions of the draft in 1997 and 2000. In 2002, EPA officials said that the assessment would conclude that dioxin may adversely affect human health at lower exposure levels than had previously been thought and that most exposure to dioxins occurs from eating such American dietary staples as meats, fish, and dairy products, which contain minute traces of dioxins. These foods contain dioxins because animals eat plants and commercial feed and drink water contaminated with dioxins, which then accumulate in animals' fatty tissue. It is clear that EPA's dioxin risk assessment could have a potentially significant impact on consumers and on the food and agriculture industries. As EPA moved closer to finalizing the assessment, in 2003 the agency was directed in a congressional appropriations conference committee report to not issue the assessment until it had been reviewed by the National Academies. The National Academies provided EPA with a report in July 2006. In developing a response to the report, which the agency is currently doing, EPA must include new studies and risk assessment approaches that did not exist when the assessment was drafted. EPA officials said the assessment will be subject to the IRIS review process once its response to the National Academies' report is drafted. As of 2008, EPA has been developing the dioxin assessment, which has potentially significant health implications for all Americans, for 17 years. While Appropriate Coordination with Federal Agencies Could Help EPA Resolve IRIS Assessment Controversies More Efficiently, EPA's Proposed Expansion of Agencies' Roles in IRIS Assessments Would Cause Further Delays and Limit Their Credibility: Although an interagency review process managed by OMB was informally incorporated into the IRIS assessment starting in 2004, federal agencies continue to believe they should have greater and more formal roles in EPA's development of IRIS assessments given the potential impact of the assessments on either their missions or their budgets, such as the need to redesign systems to eliminate hazardous materials or to clean up contaminated sites. These agencies--including DOD, DOE, the Department of Homeland Security, the Department of Transportation, and NASA--have sought earlier, more formal involvement in IRIS assessments than is currently provided through the OMB/interagency review process and the other avenues for input that currently exist, which include nominating chemicals for assessment and providing relevant studies for planned or ongoing assessments. Officials from DOD, NASA, and DOE told us there is a lack of formality and transparency about how they can provide input and when it is appropriate to do so. For example, they seek a more formal process for nominating chemicals and providing relevant studies for planned or ongoing assessments, and they want to help decide questions for the peer reviews of IRIS assessments. They said that reducing the scientific uncertainty in IRIS assessments is important because some assessments can have significant impacts on their operations and budgets. Also, the officials said that their involvement with IRIS assessments occurs too late in the process, leading to disagreement among EPA and the agencies after the assessments are drafted, and unnecessarily delaying certain assessments, such as naphthalene, TCE, and dioxin. Along these lines, we concluded in a 2006 report that EPA could help ensure consistent, transparent, and high-quality risk assessments by working with stakeholders early and periodically throughout the process to identify, among other things, key issues and studies that need to be considered in the analysis.[Footnote 37] Nonetheless, we believe that if EPA is to increase the involvement of DOD and other potentially affected federal agencies in the IRIS assessment process, it is important to do so in a way that enables EPA to balance the benefits of increased interagency coordination with EPA's need to improve the timeliness of its IRIS assessments and to ensure their credibility with proper controls. In response to the continuing concerns of some federal agencies regarding the IRIS assessment process, EPA has, for several years, been working to develop a formal IRIS assessment process that would give other federal agencies a more significant role in the process.[Footnote 38] EPA consulted with an Interagency Work Group on the IRIS Process, which includes officials representing potentially affected agencies such as DOD, DOE, and NASA, as well as officials from other agencies that develop health assessments of chemicals, such as the Department of Health and Human Services' ATSDR. OMB, the Department of Homeland Security, and the White House Office of Science and Technology Policy are also represented on the working group. EPA developed a draft process in early 2007 that the agency believed would be acceptable to the interagency work group. One of the key aspects of EPA's draft process is an increased role for other agencies- -including those whose operations and resource levels could be affected by the results of the assessments--in providing input to IRIS chemical assessments. Specifically, federal agencies would have the opportunity to be involved, or provide some form of input, at almost every step of EPA's IRIS assessment process. Most significantly, the draft proposed process would provide federal agencies with several opportunities during the IRIS assessment process to identify chemicals of interest to them as "mission critical." As defined in the draft process, a mission- critical chemical is one that is "an integral component to the successful and safe conduct of an agency's mission in any or all phases of operations." Potential impacts on the use of mission-critical chemicals assessed under the IRIS program include "cessation or degradation of the conduct of the mission and/or unacceptable resource constraints."[Footnote 39] A mission-critical designation would add requirements to the assessment process, providing other federal agencies with increased involvement in IRIS assessment decisions. As outlined, this designation could add 2 or more years to the process (see shaded boxes in fig. 5 for these additional requirements). Figure 5: EPA's Draft Proposed IRIS Assessment Process: [See PDF for image] This figure is an illustration of EPA's Draft Proposed IRIS Assessment Process, as follows: 1. Annual request for nomination of IRIS chemicals for assessment[A]. 2. Determine the annual agenda and publish it in the Federal Register[A]. 3. Scientific literature search. 4. Federal Register notice/data call-in[A]. 5. Development of a draft qualitative assessment. Is the chemical mission critical? If yes, go to step 6; if no, go to step 10. 6. Public review of draft qualitative assessment[B]. 7. Review of public comments[B]. 8. Interagency evaluation to close data gaps[B]. Is there interest in conducting research to close data gaps? If yes, go to step 9; if no, go to step 10. 9. Federal agency development of new studies[B]. 10. Development of a complete draft IRIS assessment (qualitative and quantitative), including internal peer consultation. Is the chemical mission critical? If yes, go to step 11; if no, go to step 12. 11. EPA consults with federal agencies to determine level and scope of peer review[B]. Go to step 13. 12. EPA determines the level and scope of the peer review. 13. Internal agency review by the 18-member IRIS Agency Review Committee. 14. Revised draft IRIS assessment. Return to step 13 if necessary. 15. OMB/interagency review coordinated by OMB[C]. 16. Revised draft IRIS assessment. Return to step 15 if necessary. 17. Independent external peer review and public comment[C]. 18. Revised draft IRIS assessment. 19A. Second internal agency review by the 18-member IRIS Agency Review Committee. Return to step 18 if necessary. 19B. Second OMB/interagency review coordinated by OMB[C]. Return to step 18 if necessary. 20. Completion of IRIS assessment, EPA management review and approval, and posting on IRIS. [A] Additional steps, under EPA’s planned changes, to its assessment process and indicate steps where EPA has provided additional opportunity for input from potentially affected federal agencies for mission-critical chemicals. [B] Indicate steps where EPA has provided additional opportunity for input from potentially affected federal agencies for all chemicals. [C] Indicate steps where potentially affected federal agencies already had an opportunity for input. Source: GAO analysis of EPA information. [End of figure] The additional requirements for assessments of mission-critical chemicals that would lengthen the IRIS assessment process include the following: * The opportunity for federal agencies and the public to identify any additional information that is available on a chemical and to correct any errors on EPA's new first assessment draft--a draft qualitative assessment.[Footnote 40] EPA estimates that the public comment period alone would add 45 days to the assessment process. * An opportunity for potentially affected federal agencies to review public comments made during the error correction step and initiate a meeting with EPA if they want to discuss a particular set of comments. EPA estimates that this step, which includes EPA's review and analysis of the public comments, would add 30 days to the process. * An opportunity for potentially affected federal agencies to fill a data gap or eliminate an uncertainty factor that is identified in the qualitative draft. If an agency believes it can sponsor research to fill a significant data gap, EPA would then suspend the assessment process for up to 18 months to allow that agency time to conduct research and have the completed study peer reviewed. In addition, the draft proposed process would give other federal agencies a formal role in helping EPA determine the level of independent peer review assessments for mission-critical chemicals-- that is, whether the peer reviews would be conducted by EPA Science Advisory Board panels, National Academies' panels, or panels organized by an EPA contractor. In addition, the other federal agencies would be able to help determine the panel members' areas of scientific expertise as well as the scope of the peer reviews and the specific issues they would address. Finally, the draft proposed process would also formalize the roles of other federal agencies in nominating chemicals for assessment, helping EPA determine the chemicals it would assess, and providing scientific information (reports, studies, etc.) for the assessments EPA undertakes. In the past, EPA sought such information in public notices in the Federal Register, and federal agencies could have provided their input either in response to these notices or by contacting EPA officials directly. However, in response to criticisms that EPA's approach to seeking input from the public was not appropriate for other federal agencies, under the proposed process, EPA would directly reach out to other federal agencies for this input by separately communicating these requests to them at the time it publishes its notices in the Federal Register. The draft proposed IRIS process would also formalize the OMB/interagency reviews, which provide other federal agencies with the opportunity to provide comments on IRIS draft assessments both before and after the draft assessments are provided to external peer reviewers. The OMB/interagency reviews would continue to be managed by OMB. EPA has acknowledged that the additional steps and opportunities for input from other federal agencies that its draft proposed process would provide will add more time to an already lengthy process. Specifically, under the proposed process we reviewed, EPA officials estimated that IRIS assessments for standard chemicals would take roughly 2-1/2 years to complete and that an additional 2 to 3 years would be needed for mission-critical chemicals. In February 2008, the Assistant Administrator, Office of Research and Development, updated these time frames, estimating that most assessments would take between 3 and 4-1/ 2 years to complete, while assessments of mission-critical chemicals would take up to 6 years. However, these estimates appear to be optimistic considering the length of time ongoing assessments--which have not been subject to additional process steps--have been in progress: most for more than 5 years and many for more than 9 years. Further, when assessments take longer than 2 years, they can become subject to substantial delays stemming from the need to redo key analyses to take into account changing science and assessment methodologies. According to officials from EPA and some of the potentially affected federal agencies, the key goals of the draft proposed changes to the IRIS assessment process are to (1) fill data gaps to reduce uncertainty in the IRIS assessments and (2) prevent disagreements among EPA and other federal agencies about assessments late in the IRIS assessment process. As discussed earlier, disagreements have arisen late in the process in the past and have significantly delayed the completion of some assessments. These goals are important, and it is appropriate for EPA and the federal agencies to find ways to achieve them. However, some of the proposed changes would result in assessments being caught in a continuous cycle of updates and revisions--delaying the completion of assessments of chemicals with potentially significant health impacts for many years. As discussed above, delays for any reason--including waiting for new research to fill data gaps--can have a compounding effect on time frames because delays often require extensive rework, such as updates to reflect other studies that have come out in the interim, performing additional analyses to reflect the current state of the science of risk assessment tools, and other rounds of interagency or external peer reviews because of changes in the assessments. Thus, changing from a reliance on the best available scientific data at the time an assessment is conducted to a reliance on research that has either not yet started or been completed in order to fill scientific data gaps is a decision that can have a significant impact on assessment completion dates. Further, although EPA has said that IRIS assessments would only be suspended to wait for shorter-term studies, such as those assessing modes of action, we and several agency officials we spoke with believe that the time needed to plan, conduct, and complete research that would address significant data gaps, and have it peer reviewed, would likely take longer than the 18 months EPA would allow under its draft proposed IRIS process. In addition, although the draft process would set a limit on the time the assessments can be suspended to await studies, the enforcement of such time frames could prove to be difficult when research schedules slip. And not waiting for study results beyond the 18-month limit could, in fact, undo all of the efforts expended to avoid disagreements in the later stages of the assessments. Also, EPA's prior experience with waiting for studies, such as in the case of the formaldehyde assessment discussed earlier, shows that estimates of completion dates for these studies can stretch out from months to years. Finally, the extent to which new research would provide information that would actually reduce key scientific uncertainties in the IRIS assessments is not known until the research is completed and peer reviewed. In discussing with us the need for additional time to conduct new research to fill scientific data gaps for chemicals defined as mission critical, DOD officials emphasized that the federal agencies believe it is important that scientific uncertainties are reduced to the maximum extent practical within a reasonable time frame before IRIS assessments of mission-critical chemicals with high uncertainty are completed. The potential impacts of IRIS assessments on DOD--should risk assessments using the IRIS data lead to regulatory actions--could include the need to (1) conduct research and development of material substitutes; (2) redesign systems and processes to eliminate hazardous materials; (3) conduct research and development of treatment or cleanup technologies; (4) improve personal protective clothing, equipment, and procedures; and (5) provide for special handling and storage of chemicals. However, the reality is that scientific data gaps will almost always exist. EPA, whose mission includes protecting the public and the environment from harmful chemicals, must decide whether to use assumptions and estimates to address data gaps or to wait for research that could potentially fill some of the data gaps. On this issue, in discussing the fact that EPA must address many environmental risks with incomplete data and a lack of consensus about assumptions, the agency has stated that it seeks to strike a balance among fairness, accuracy, and efficient action because not acting until data quality improves can result in substantial harm to human health, safety, and the environment. Along these lines, as discussed earlier, a National Academies panel recently reported on a chemical of interest (trichloroethylene) to several federal agencies that would likely be defined as mission critical under the proposed IRIS process. The National Academies' report recommended that the risk assessment be finalized with currently available data, rather than waiting for additional information to address scientific uncertainties, so that risk management decisions can be made expeditiously because the evidence on risks to human health have strengthened. The National Academies specifically noted that while some additional information would allow for more precise estimates of risk, this information is not necessary at this time for developing a credible risk assessment. EPA started the assessment in 1998, and it has been subject to multiple independent peer reviews. However, EPA is currently preparing another assessment draft that will need to undergo the IRIS assessment process a second time. As such, the draft will be to subject to the OMB/ interagency reviews and another independent peer review before it can be finalized. We note that if EPA's proposed IRIS process is implemented, DOD and other federal agencies could designate trichloroethylene as a mission-critical chemical and have the assessment suspended for 18 months for research aimed at reducing scientific uncertainties. EPA could potentially minimize or eliminate requests to suspend ongoing assessments to fill in scientific data gaps by better coordinating with DOD's emerging contaminants program--established in 2006--that leads and supports the Materials of Emerging Regulatory Interest Team (MERIT).[Footnote 41] The intent of the emerging contaminants program is to help DOD proactively manage chemicals DOD uses or has used in its operations that (1) are a perceived or real threat to human health or the environment and (2) lack a published health standard and/or have new information available about their sources or effects. The program seeks to identify chemicals of importance to DOD's mission that are on the regulatory horizon so that the agency can begin to make changes or identify areas of research that may minimize the impact that regulations of a particular chemical would have on the agency and its mission. Of particular relevance to the IRIS program is the scanning component of the emerging contaminants program, which consists of a regular review of the scientific literature, periodicals, and regulatory communications for chemicals that have, or are likely to have, changing human health values or regulatory standards. DOD then further screens those chemicals that are, were, or will be used by DOD to determine whether a potential impact to DOD's mission exists. After screening, DOD conducts impact assessments to determine the severity of risk to DOD functional areas, which include environment, safety, and health protection; research and development; production, operation and maintenance, and disposal; training and readiness; and cleanup. Through these efforts, DOD seeks to identify cost-effective opportunities to proactively manage future unacceptable risks. One research area the emerging contaminants program has identified as a potentially good investment is DOD-funded health effects research aimed at filling important data gaps for mission-critical chemicals with high uncertainty factors.[Footnote 42] DOD's ongoing actions in monitoring scientific literature and identifying research that could reduce scientific uncertainties could be more effectively put to use. Specifically, enhanced coordination and collaboration between EPA and DOD's emerging contaminants program have the potential to proactively reduce some scientific uncertainties associated with chemicals that EPA is not yet assessing--as opposed to the current reactive focus on ongoing IRIS assessments. That is, under the draft proposed IRIS process changes we reviewed, DOD and other federal agencies would be able to designate ongoing chemicals as mission critical at several stages in the assessment process--and have their assessments suspended for up to 18 months to address data gaps. However, enhanced coordination and collaboration between DOD and EPA could change this essentially reactive approach to a model that is more proactive and effective. If, for example, EPA provided 2 years' notice of its intent to assess specific chemicals, rather than simply announcing the list of chemical assessments it plans to initiate during a given fiscal year, DOD and other agencies could use this time to sponsor and complete the studies they deem necessary, rather than waiting to initiate them after an IRIS assessment had started. Giving agencies 2 years' notice (which would provide them with more than 18 months to conduct additional research) would be more effective and efficient than delaying ongoing IRIS assessments for 18 months to await research, given the compounding effect of delays on IRIS assessments. We note that, to date, the emerging contaminants action list--the list of chemicals that DOD has assessed and judged to have a significant potential impact on people or the DOD mission--has focused largely on chemicals that EPA's IRIS program is already assessing. However, with some adjustments in the emerging contaminants program's focus or scope, DOD could widen its scope and help EPA with its planning process by earlier identification of the mission-critical chemicals of concern to DOD for which IRIS assessments are needed. Finally, when the emerging contaminants program identifies health research needs for chemicals considered mission critical, DOD could meet its research needs more effectively and quickly by sponsoring research as needs are identified, rather than waiting until IRIS assessments are started. Lastly, we note that while increased coordination with potentially affected federal agencies on IRIS assessments could enhance the assessments and facilitate their completion, giving potentially affected federal agencies formal roles in some IRIS assessment decisions reduces the credibility of the assessments if proper controls, such as ensuring transparency, are not in place. While we recommended in our 2006 report on human health risk assessment that EPA consistently involve stakeholders as appropriate to the risk assessment, we made this recommendation in the context of improving the overall quality, consistency, and transparency of risk assessments. However, one aspect of EPA's draft proposed IRIS process that has proven to be controversial involves transparency. Specifically, EPA's early 2007 draft proposal included making the comments from OMB and other federal agencies provided during the OMB/interagency review process part of the public record; according to EPA, this requirement has been removed from the latest draft proposal because of concerns raised by OMB. These concerns have delayed the implementation of the planned process changes, and as of December 2007, EPA was planning to send a revised draft back to the interagency work group for review and approval. According to DOD officials, OMB did not support certain aspects of the draft process EPA presented to the Interagency Work Group on the IRIS Process in early 2007, expressing concerns about how the proposal would address the concerns of states, industry, and environmental entities and the requirement to have agencies' comments part of the public record. The latter concern stemmed from the precedent it might set and because OMB believes that doing so would erode the effectiveness of the deliberative process among EPA and the agencies. However, because the agencies' comments on IRIS assessments are to be scientific in nature, it is unclear why agencies might not candidly comment on an IRIS assessment even if their comments would become part of the public record. In fact, officials from DOD, DOE, and NASA told us they did not object to EPA's plan to make interagency comments part of the public record because they would be scientific in nature and not part of a policy discussion. While OMB officials would not comment directly on this issue as it relates to objections they might have to EPA's draft proposed process, one official told us that generally OMB believes that effective deliberations among federal agencies are important and that if agencies' deliberative comments are part of the public record, agency officials will not be as frank and candid as they would be under the protection of confidentiality. However, transparency in the IRIS assessment process can provide assurance that these scientific assessments are appropriately based on the best available science and that they are not impacted by policy issues and considerations. Under the National Academies' risk assessment and risk management paradigm, policy considerations are relevant in the risk management phase, which occurs after the risk assessment phase that encompasses IRIS assessments.[Footnote 43] Some of the federal agencies that would be given formal roles in some IRIS assessment decisions are particularly interested in risk management issues as they would likely face adverse consequences, such as increased cleanup costs and other legal liabilities, if EPA issues an IRIS assessment for a given chemical that results in a risk management decision to regulate the chemical to protect the public.[Footnote 44] Consequently, to ensure credibility of the IRIS assessments, the input of these agencies into the assessment process should be transparent. Further, when EPA and other agencies propose regulations or other actions on the basis of health risk assessments, the proposals represent policy choices. Through the rule-making process and other interagency working groups, federal agencies have the opportunity to participate in policy dialogues. Such discussions can appropriately address risk management concerns that, by definition, involve integrating risk characterization information (based, in part, on information in IRIS) with other information[Footnote 45] to decide how to protect public health. In contrast, the input of federal agencies into the IRIS assessments, part of the risk assessment process, should be based solely on science issues, not policy concerns. Conclusions: The IRIS database, one of the most significant tools that EPA has developed to effectively support its efforts to protect people and the environment from harmful chemical exposures, is at serious risk of becoming obsolete because the agency has not been able to keep its existing assessments current or complete assessments of the most important chemicals of concern. Although EPA has taken important steps to improve the IRIS program and productivity since 2000 and has developed a number of draft assessments for external review, its efforts to finalize the assessments have been thwarted by a combination of factors: the imposition of external requirements, the growing complexity and scope of risk assessments, and certain EPA management decisions. Each of these factors has led to delays in the completion of individual IRIS assessments. In addition, an overarching factor--the compounding effect of delays--has had a particularly profound impact on productivity: Even a single delay can create a cascading series of delays with far-reaching, time-consuming consequences. In fact, in some cases, it is necessary to essentially start the assessment process over because of the need to incorporate science and methodologies that have evolved since the assessments began. To effectively meet its diverse user needs, EPA must both keep the IRIS database up to date and undertake assessments of potentially dangerous chemicals not yet evaluated. Achieving these objectives will require EPA to complete IRIS assessments in a timely fashion that minimizes rework--an outcome that cannot occur unless EPA's IRIS assessment process is streamlined to routinely support timely completion of assessments. When assessments take longer than two years, they can become subject to substantial delays stemming from the need to redo key analyses to take into account changing science and assessment methodologies. However, the manner in which other federal agencies are involved in the IRIS assessment process prevents EPA from effectively streamlining and managing its assessments. Under the current process, EPA is required to send draft assessments to OMB at two key points in the assessment process. Although the OMB/interagency reviews have no time frames or deadlines, EPA is not allowed to proceed with the assessment process until OMB agrees that EPA has sufficiently responded to agencies' comments; and OMB has also directed EPA to terminate five assessments. Further, OMB's view that disagreements between EPA and other agencies should simply be elevated to senior EPA executives for resolution is inefficient and causes delays. That is, an effective IRIS assessment process would not require the constant intervention of top-level executives, whose time is already at a premium. Moreover, the current process elevates the goal of reaching interagency agreement above achievement of IRIS program objectives and, in doing so, fails to adequately acknowledge the expertise of EPA scientists and the many other scientific experts who have prepared and reviewed the assessments. We believe that interagency coordination can enhance the quality of EPA's IRIS assessments. However, this goal would more likely be achieved if the input from other federal agencies was obtained in a manner that better ensured that EPA's scientific analyses were given appropriate weight and that provided time frames to avoid indefinite delays. Because an integral part of EPA's mission is to assess the risks associated with exposures to chemicals, the agency is best situated to establish how--and at what level--to best resolve IRIS assessment issues raised by other federal agencies. Until EPA can establish time frames for various assessment steps--including the OMB/ interagency review process--and determine (1) how it will resolve interagency conflicts, (2) when assessments are ready for either independent peer review or completion, and (3) the scope of the assessments needed to support EPA programs, EPA will continue to be hampered in its ability to develop timely and credible IRIS assessments that meet its needs and protect the environment. Importantly, the lack of transparency of the OMB/interagency review process reduces the credibility of EPA's IRIS assessments. Because the agencies' comments and the changes EPA makes in response are treated as internal executive branch documents not subject to release outside the executive branch, the OMB/interagency reviews occur in what amounts to a black box. Specifically, the first assessment drafts that become publicly available--those provided to independent peer reviews-- incorporate changes from federal agencies that may be affected by the assessments without identifying or providing reasons for the changes. Moreover, the changes EPA is proposing to its current assessment process would provide potentially affected federal agencies with a formalized opportunity to provide input at almost every step in the assessment process without addressing the extent to which agencies' input would be made part of the public record. Given the importance and sensitivity of IRIS assessments, it is critical that input from all parties, particularly agencies that may be affected by the outcome, be publicly available to alleviate concerns of potential bias in the assessments. In addition, EPA's draft proposed changes to its IRIS assessment process--which would further expand the role of other federal agencies in the process, including the option to suspend assessments of key chemicals for up to 18 months to develop new studies--would add to the already unacceptable level of delays in completing IRIS assessments. Under its draft proposed changes, EPA estimates that assessments of certain key chemicals may take up to 6 years. Assessments of this duration are problematic--they tend to get caught in a perpetual cycle of updates and revisions. Further, we believe EPA's time estimate for assessments under its draft proposal are optimistic, as most IRIS assessments currently in progress have already been in the works for more than 5 years, including 12 that have been in progress for more than 9 years. Yet these assessments have not been subject to the additional steps and requirements under the draft proposed process. An alternative, more efficient approach to suspending assessments while waiting for new research would be for EPA to give agencies and the public more advance notice of planned assessments, thereby providing external parties with an interest in conducting additional research on a given chemical the ability to complete their work prior to the start of the IRIS assessment. This is important because effectively maintaining the IRIS database depends on strict adherence to time frames using the best available science. In addition, we note that OMB has raised questions about whether the proposed process sufficiently addresses the needs of private entities. Along these lines, we believe that if EPA grants special rights to other federal agencies to suspend IRIS assessments to conduct new research, it is likely that the agency would face pressure to do so for numerous other entities, including industries and individual companies that could be impacted by IRIS assessments should the assessments lead to regulatory actions. Further, while we believe it is appropriate for EPA to identify key uncertainties in IRIS assessments and to continue to work on developing methodologies for conducting more sophisticated and meaningful uncertainty analyses, continuing to delay IRIS assessments in order to develop and test enhanced methodologies and strategies for communicating them can conflict with EPA's stated goal of seeking balance among fairness, accuracy, and efficient action. More specifically, EPA has acknowledged that not acting until data quality improves can result in substantial harm to human health, safety, and the environment. Also, as EPA continues to incorporate uncertainty analysis in its assessments, ensuring that the information is clear and useful to decision makers is important. Along these lines, the National Academies and others have warned that producing "ranges of meaningless and confusing risk estimates" could result in assessments of reduced, rather than enhanced, quality and objectivity. Lastly, while it is difficult to overstate the importance of the IRIS program to EPA's ability to effectively conduct its mission of protecting human health and the environment, this program currently uses about 0.1 percent of EPA's annual appropriations--specifically, in fiscal year 2007, the program received about $9.6 million of EPA's $7.3 billion budget. EPA's current estimate that it will be able to complete 16 assessments a year by 2011 would represent a substantial increase over recent productivity; however, it is not clear that this level will be sufficient to maintain the viability of the IRIS database. Recommendations for Executive Action: To develop timely chemical risk information that EPA needs to effectively conduct its mission, we are recommending that the Administrator, EPA, require the Office of Research and Development to re-evaluate its draft proposed changes to the IRIS assessment process in light of the issues raised in this report and ensure that any revised process: * clearly defines and documents a streamlined IRIS assessment process that can be conducted within time frames that minimize the need for wasteful rework and carefully considers the trade-off between the benefits of changes that involve additional steps and time--including the development of enhanced uncertainty analyses and other emerging assessment methods--and the impacts of these changes on EPA's ability to complete timely chemical assessments; * establishes a policy that endorses conducting IRIS assessments on the basis of peer-reviewed scientific studies available at the time of the assessment and develops criteria for allowing assessments to be suspended to await the completion of scientific studies only under exceptional circumstances; * establishes IRIS assessment needs to provide at least 2 years' notice of assessments that are planned, including criteria for making exceptions to the advance notifications, if needed; * sets time limits for all parties, including OMB and other federal agencies, to provide comments to EPA on draft IRIS assessments; and: * periodically assesses the level of resources that should be dedicated to this significant program to meet user needs and maintain a viable IRIS database. In addition, we recommend that the EPA Administrator take steps to better ensure that EPA has the ability to develop transparent, credible IRIS chemical assessments--an ability that relies in large part on EPA's independence in conducting these important assessments. Actions that are key to this ability include ensuring that EPA: * can determine the types of IRIS assessments to conduct on the basis of the needs of EPA's program offices and other users; * can define the appropriate role of external federal agencies in EPA's IRIS assessment process and manage an interagency review process in a manner that enhances the quality, transparency, timeliness, and credibility of IRIS assessments, including determining when interagency issues have been appropriately addressed; and: * has the ability to provide comments by OMB and other federal agencies on draft IRIS assessments to decision makers, the Congress, and the public. Agency Comments and Our Evaluation: In commenting on a draft of this report, EPA's Assistant Administrator for Research and Development agreed to consider our recommendations in revising the IRIS assessment process. However, EPA stated that it believed the productivity and transparency issues discussed in the draft report were misrepresented in the title and body of the report. We disagree and believe we have fairly represented IRIS productivity and transparency issues related to the IRIS assessment process. We did clarify that the transparency issues highlighted in our report focus on the IRIS assessment process rather than on the content of IRIS assessments, and we revised the report title. In addition, EPA emphasized that the proposed changes to the IRIS assessment process are still subject to change. Along these lines, we made our recommendations with the intent that they would be integrated into any revised process. In its response, EPA also estimated that under the new process, most assessments would take between 3 and 4-1/2 years; mission-critical assessments would take up to 6 years. These estimates differ from the time frames EPA officials provided during our review, and we have revised the report to reflect this. However, we believe an IRIS assessment process built around such time frames is problematic. As we state in our draft and final reports, when assessments take longer than 2 years, they can become subject to substantial delays stemming from the need to redo key analyses to take into account changing science and assessment methodologies. Finally, EPA asserted that the new process will lead to increased productivity. We disagree and have concluded that the proposed revisions would add to the already unacceptable level of delays in completing IRIS assessments. In its comments, OMB did not specifically address the recommendations we made to EPA but disagreed with some aspects of the report, primarily regarding our characterization of the OMB-managed interagency reviews and their effects on IRIS assessments. For example, OMB indicated that it disagreed with our conclusions that the OMB/interagency reviews make it more difficult for EPA to complete IRIS assessments in a timely manner, that these reviews affect the credibility of the assessments, and that interagency comments should be transparent. We disagree with OMB and believe that we have fairly represented the OMB/interagency review process as well as the importance of input from all parties being publicly available. Given the importance and sensitivity of IRIS assessments, it is critical that input from all parties, particularly agencies that may be affected by the outcome, be publicly available to alleviate concerns of potential bias in the assessments. EPA's and OMB's comments and our detailed responses appear in appendix IV and appendix V. As agreed with your office, unless you publicly announce the contents of this report earlier, we plan no further distribution of this report until 30 days from the report date. At that time, we will send copies to the Administrator, EPA; the Director, OMB; and appropriate congressional committees and other interested parties. We will also make copies available to others on request. In addition, the report will be available at no charge on the GAO Web site at [hyperlink, http://www.gao.gov]. If you or your staff have any questions about this report, please contact me at (202) 512-3841 or stephensonj@gao.gov. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff who made major contributions to this report are listed in appendix VI. Sincerely yours, Signed by: John B. Stephenson: Director, Natural Resources and Environment: [End of section] Appendix I: Objectives, Scope, and Methodology: This appendix details the methods we used to assess the Environmental Protection Agency's (EPA) management of its Integrated Risk Information System (IRIS). For this review, we determined (1) the outcome of steps that EPA has taken to ensure that IRIS contains current, credible chemical risk information, to address the backlog of IRIS assessments, and to respond to new requirements from the Office of Management and Budget (OMB); and (2) the potential effects of EPA's planned changes to the IRIS assessment process on EPA's ability to ensure that IRIS provides current, credible risk information. To address these two objectives, we reviewed relevant EPA and other documents, including EPA's 2003 IRIS Needs Assessment, the Office of Research and Development's multiyear plans, numerous documents outlining the IRIS assessment process and draft proposed changes to the assessment process, EPA budget justification documents, and OMB's Program Assessment Rating Tool assessment that covered the IRIS program. We interviewed officials from EPA's National Center for Environmental Assessment who manage the IRIS assessment program, including the Center Director, the Associate Director for Health, and the IRIS Program Director; officials from other EPA program offices, including the Office of Air and Radiation; and officials from federal science and health agencies that are involved in the IRIS assessment process, to obtain their perspectives on, among other things, the current IRIS assessment process, the potential effects of the draft proposed changes to the process, the extent to which EPA has made progress in completing assessments and meeting user needs, and challenges EPA faces in completing assessments. In addition, we interviewed officials from the Department of Defense, the Department of Energy, and the National Aeronautics and Space Administration who have served on the interagency working group on the IRIS assessment process, and OMB officials in the Office of General Counsel and the Office of Information and Regulatory Affairs to obtain their perspectives on the OMB/interagency review process and on the planned changes to the IRIS assessment process. We also interviewed officials from the Department of Health and Human Services' Agency for Toxic Substances and Disease Registry (ATSDR), an agency that assesses the potential noncancer health effects of exposure to some chemicals, to obtain information on ATSDR's assessment process. In addition, we attended the Board of Scientific Counselors Human Health Risk Assessment (HHRA) Subcommittee meetings in November 2007. During these meetings, the Subcommittee reviewed the EPA Office of Research and Development's HHRA program-- specifically, its relevance, quality, performance, and scientific leadership. For the first objective, we obtained and analyzed several data sets, including data from EPA's IRIS database; EPA's IRIS Track--a compilation of status reports for IRIS assessments in progress; Federal Register notices announcing annual IRIS agendas; a spreadsheet that EPA uses to track the IRIS assessments sent to OMB for the OMB/interagency reviews; EPA's 2003 screening level reviews report--its efforts to identify assessments that may need to be updated; and EPA's 2003 IRIS Needs Assessment. From the data we gathered, we analyzed information on EPA's productivity, including the number of new and completed IRIS assessments, for fiscal years 2000 through 2007; the status of IRIS assessments, as of December 1, 2007, that were in progress during fiscal year 2007; the status of IRIS assessments that have been sent to OMB for OMB/interagency review and the time elapsed during those reviews; the number of assessments in the IRIS database that may need to be updated; the resources provided to the program for fiscal years 2000 through 2007; and user needs and EPA's assessment completion goals. We also interviewed chemical managers, who are responsible for managing the IRIS assessments in progress, to obtain further information on the 77 chemicals in progress during fiscal year 2007. We did not evaluate the scientific content or quality of IRIS assessments. In addition, we conducted a reliability review of the data we received from EPA for our first objective. Through our review, we determined that the data we used were sufficiently reliable for our purposes. Our assessment consisted of interviews with officials about the data system and elements and the method of data input, among other areas. We also corroborated the data with other sources, where possible. For example, we verified the information provided in EPA's IRIS Track system with the chemical managers responsible for each of the 77 chemicals in progress during fiscal year 2007, and updated the information as appropriate. Similarly, we ensured that status information presented in IRIS Track was consistent with the status information presented in the spreadsheet that EPA uses to track draft assessments provided to OMB for the OMB/interagency review. In addition, EPA attempted to corroborate its data on the status of IRIS assessments that have been sent to OMB for OMB/interagency review and the time elapsed during those reviews, providing OMB with its tracking spreadsheet for review. OMB chose not to respond. Consequently, we relied on EPA's data and assessed its reliability, based on information including the source of the data and method of input. We conducted this performance audit from October 2006 to March 2008 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. [End of section] Appendix II: Broad Outline of EPA's Current IRIS Assessment Process: Note: This outline does not include the individual activities, including internal reviews and briefings, encompassed within many of the broad IRIS assessment categories. 1. Annual request for nomination of IRIS chemicals for assessment. 2. Determine the annual agenda and publish it in the Federal Register, along with a call for scientific information from the public on select chemicals. 3. Scientific literature search. 4. Development of a complete draft IRIS assessment (qualitative and quantitative), including internal peer consultation. 5. Internal agency review by the 18-member IRIS Agency Review Committee. 6. Revised draft IRIS assessment. 7. OMB/interagency review coordinated by OMB. 8. Revised draft IRIS assessment. 9. Independent external peer review and public comment. 10. Revised draft IRIS assessment. 11A. Second internal agency review by the 18-member IRIS Agency Review Committee. Return to step 10 if necessary. 11B. Second OMB/interagency review coordinated by OMB. Return to step 10 if necessary. 12. Completion of IRIS assessment, EPA management review and approval, and posting on IRIS. Source: GAO analysis of EPA information. Note: This outline does not include the individual activities, including internal reviews and briefings, encompassed within many of the broad IRIS assessment categories. [End of section] Appendix III: Information on OMB/Interagency Reviews of IRIS Assessments: As of December 1, 2007, EPA has sent 36 draft IRIS assessments to OMB for the first OMB/interagency review managed by OMB, and the reviews of 16 assessments were completed. As shown in figure 6, the OMB/ interagency review of 4 assessments was completed in less than 6 months and 12 of these assessments added 6 or more months to the IRIS assessment process. Figure 6: Status of IRIS Assessments Sent to OMB for the First OMB/ Interagency Review Starting in Fiscal Year 2004, as of December 1, 2007: [See PDF for image] This figure is a vertical bar graph depicting the following data: Number of assessments: Status: Completed interagency review in less than 6 months; Number of assessments: 4. Status: Completed interagency review in 6 or more months; Number of assessments: 12. Status: Assessment suspended, now being redone; Number of assessments: 1. Status: Assessment discontinued after a year in OMB/interagency review; Number of assessments: 5. Status: OMB/Interagency review ongoing for 6 months to a year; Number of assessments: 1. Status: "Interagency review ongoing for 2 to 4 months[A]; Number of assessments: 13. Source: GAO analysis of EPA data. [A] These 13 assessments were submitted to OMB in the last quarter of fiscal year 2007. [End of figure] Moreover, as also shown in figure 6, five assessments that assessed acute exposure were discontinued at the direction of OMB after a year in the OMB/interagency review stage,[Footnote 46] and one assessment was suspended and is now being redone (naphthalene). Finally, 14 assessments were at the OMB/interagency review stage as of December 1, 2007: Of these, 1 had been at this stage for 9-½ months, and the remaining 13 had been submitted to OMB in the last quarter of fiscal year 2007 and had been at the OMB/interagency review stage for 2 to 4 months. While OMB officials said that EPA does not have to provide a document addressing every OMB/interagency comment, EPA officials said that a detailed disposition of comments document was, in fact, necessary in order to get OMB's agreement that EPA had satisfactorily addressed all comments. There are no time frames for this iterative process. OMB officials said that the second OMB/interagency review is conducted by OMB to ensure that EPA has adequately considered the comments from the external peer review panel. As of December 1, 2007, 10 assessments had been sent to OMB for the second OMB/interagency review, and 5 of these assessments completed the review. The time frames for these five assessments ranged from 10 days to almost 4 months. [End of section] Appendix IV: Comments from the Environmental Protection Agency: Note: GAO comments supplementing those in the report text appear at the end of this appendix. United States Environmental Protection Agency: Office Of Research And Development: Washington, D.C. 20460: [hyperlink, http://www.epa.gov]: February 21, 2008 John B. Stephenson, Director: Natural Resources and Environment: Government Accountability Office: 441 G Street NW, Room 2135: Washington, DC 20548: Dear Mr. Stephenson: Thank you for the opportunity to review and comment on the Government Accountability Office's (GAO) draft report, "Low Productivity and Lack of Transparency Limit the Usefulness and Credibility of EPA's Integrated Risk Information System" (IRIS). Recognizing that EPA will, in all probability, respond to the final version, I would like to comment briefly on the productivity and transparency issues that are discussed in the draft report, which we believe are misrepresented in the report's title and throughout the report. Since GAO raised many issues already recognized by EPA, the Agency has been actively working on revisions to its IRIS process. We believe these revisions will improve the productivity and transparency of IRIS. It should be noted that although GAO discusses some of the aspects of EPA's ongoing efforts to update the IRIS process, and comments on the potential effects of the proposed changes to the process, the GAO draft report refers to an outdated draft (internal) document. [See comment 1] Consequently, some of the steps or procedures presented by GAO in its draft report have changed or will no longer be relevant in the final process. GAO's draft report should acknowledge that they were working from a draft internal document, and therefore that GAO's conclusions and recommendations may be erroneous when compared to the final IRIS process. [See comment 2] Nonetheless, EPA will consider each of the GAO's recommendations in light of the new IRIS process, even though these recommendations were based on the outdated draft document. Under the new process, EPA's increased involvement of other agencies and the public will help us identify scientific issues earlier and foster better communication and sharing of information, which will ultimately help streamline the IRIS process. [See comment 3] It is worth noting that under the new process, specific time limits for each step in the process will improve the timeliness of completed assessments. For example, public and interagency review periods will be for fixed and relatively short durations. Interagency and public review will therefore represent only a small fraction of the total time required to complete an IRIS assessment. For most chemicals, the IRIS process will take between three and four and a half years. For the small number of mission-critical chemicals, the process may take an extra one to one and a half years. EPA believes the new IRIS process, in combination with the increased resources for the program provided in recent years, will lead to increased productivity in achieving completed assessments. [See comment 4] Although only briefly recognized in GAO's draft report, today's IRIS assessments are much more sophisticated, complex, and of higher scientific quality than at any other time in the program's history. [See comment 5] The increased complexity of chemical assessments often leads to extensive public comments and more rigorous levels of internal and external peer review, which is one of the most important reasons for delays in the IRIS process. Additionally, recent external peer reviews by EPA's Science Advisory Board (SAB) and the National Academy of Sciences (NAS) have requested that even more sophisticated analyses, such as quantitative uncertainty analysis, be performed and presented in some IRIS assessments. These issues, and EPA's responsiveness to them, are given too little attention or are unfairly criticized in the GAO draft report. [See comment 6] Also, the IRIS program has increased the rigor of its scientific peer review. Specifically, the external peer review process has been changed from letter reviews to face-to-face panel peer reviews that are open to the public. This has encouraged public interaction with the peer reviewers. Members of our Board of Scientific Counselors (BOSC) that recently reviewed the IRIS program indicated that "the extent of peer review for IRIS...exceeds most other examples with which the subcommittee members were familiar." [See comment 7] Transparency is critical to an effective IRIS process. It is equally important that scientists and policymakers be able to have full and frank discussions. Clearly, the quality of government decisions is better and stronger when healthy skepticism is not discouraged. [See comment 8] The current IRIS process strikes a balance between these two principles by encouraging open discussion of science and science policy questions, while appropriately protecting the deliberative process. Enhancing transparency through these and other measures necessarily means bringing many parties into the IRIS process. However, increased participation does not diminish EPA's accountability as the final decision-maker in establishing IRIS health values. All final IRIS assessments must also stand on their own merit by undergoing public and external peer review. It is noteworthy that in addition to having a transparent process, it is important to ensure transparency in the content of IRIS assessments (e.g., the rationale for relying on or using specific data sets, assumptions, or models). The GAO draft report does not comment on this facet of transparency, which directly bears on the content and quality of IRIS assessments. [See comment 9] We have taken a number of specific steps to enhance transparency in the IRIS program. For example: * We created IRIS Track on our web site to let everyone know the status of every chemical being worked on by EPA. This step opened up the process to all stakeholders, provided much greater transparency than before, and enabled interested parties to follow the progress of an assessment. [See comment 10] * Final documents are now published on the IRIS public web site. This includes all external public comments received and all the responses to the comments. The entire record of the assessment available to the public. [See comment 11] * IRIS has expanded the nomination process for new assessments to include a Federal Register Notice to allow the public to nominate chemicals, as well as a direct request to EPA Offices and other federal agencies asking for their nominations as well. [See comment 12] Like GAO, the Environmental Protection Agency highly values the IRIS Program, which stands as a model for: the international scientific community of unbiased, public, external peer review; the application of new scientific advancements in risk assessment; and focusing government efforts on priority chemical risk issues. According to the most recent review of the IRIS program by BOSC, "The comprehensiveness, transparency, and consistency of the IRIS approach has made it into the internationally-recognized standard in hazard characterization." EPA is committed to continual enhancement of the IRIS Program, and in that spirit, we appreciate your attention to the program. If you have any comments or questions on our response, please contact me or Kevin Teichman, Deputy Assistant Administrator for Science, at (202) 564- 6620. Best regards, Signed by: George Gray: Assistant Administrator: The following are GAO's comments on the Environmental Protection Agency's letter dated February 21, 2008. GAO Comments: 1. Our analysis of the draft proposed changes to the IRIS assessment process does not support EPA's assertion that the revised process EPA is developing will improve the productivity and transparency of IRIS. In fact, as discussed in our draft and final reports, the draft proposed process would exacerbate existing productivity and transparency concerns. 2. EPA emphasizes that our draft report includes information on proposed revisions to the IRIS assessment process that "have changed or will no longer be relevant" in the final version. Our draft and final reports clearly indicate that we are providing information on EPA's current assessment process as well as on proposed changes to the IRIS process that EPA has been developing for years--but has not yet finalized. For example, the draft and final reports state that "as of December 2007, EPA was planning to send a revised draft back to the interagency work group for review and approval." Moreover, in the draft and final reports, the graphic outlining the proposed changes is titled "EPA's Draft Proposed IRIS Assessment Process." Further, we made our recommendations with the intent that they be considered and integrated into any revised process that EPA finalizes. Finally, we note that EPA did not identify any significant changes it has made to the proposed process we outline in the draft and final reports, either when it commented on the draft report or on a December 2007 statement of facts provided to EPA for review for factual accuracy. However, to ensure clarity, we have added "draft" prior to our use of the term "proposed process" in a number of places. 3. EPA states that "under the new process," the increased early involvement of other agencies and the public in identifying scientific issues and sharing information will ultimately help streamline the IRIS process. We note that other agencies and the public have the opportunity for early involvement in IRIS assessments under the current process. We do not believe that simply formalizing their involvement will, in fact, streamline the process or improve the timeliness of completed assessments. 4. We have revised the report to add the updated estimated time frames for IRIS assessments under EPA's proposed revisions to the process provided in EPA's letter. With most assessments estimated to take up to 4-1/2 years to complete, it appears unlikely that the IRIS program will be able to produce quality assessments in a timely manner. As discussed in our draft and final reports, when assessments take longer than 2 years, they can become subject to substantial delays stemming from the need to redo key analyses to take into account changing science and assessment methodologies. Further, although EPA states that only a small number of IRIS assessments will be designated as mission- critical, we note that the chemicals with this designation are likely to be those posing widespread public health concerns that need to be addressed expeditiously. As we state in the draft and final reports, effectively maintaining the IRIS database will depend on strict adherence to time frames using the best available science. In addition, we believe that if EPA grants special rights to other federal agencies to, among other things, suspend IRIS assessments to conduct new research, it is likely that the agency would face pressure to do so for other entities, including industries and individual companies that could be impacted by IRIS assessments should they lead to regulatory actions. Therefore, while EPA's stated intent is to limit the number of mission-critical designations made by other federal agencies, the number is likely to increase over time. 5. While EPA states that the draft report only briefly recognizes that IRIS assessments are more sophisticated and complex, our draft and final reports identify the growing complexity of risk assessments and risk assessment methods and models as one of five key factors contributing to EPA's inability to complete IRIS assessments in a timely manner. 6. EPA states that external peer reviews have requested that more sophisticated analyses, such as quantitative uncertainty analysis, be performed and presented in some IRIS assessments. EPA further states that its responsiveness to these requests is given too little attention or is unfairly criticized in the draft report. In fact, our draft and final reports clearly state that peer reviews of EPA's assessments have sometimes recommended additional uncertainty analysis. Further, while we conclude that it is appropriate for EPA to identify key uncertainties in IRIS assessments and to continue to work on developing methodologies for conducting more sophisticated and meaningful uncertainty analyses, the methodologies are not yet developed. That is, our draft and final reports indicate that EPA plans to release draft reports in 2010 on methods for analyzing and characterizing uncertainty in hazard and dose response and in physiologically based pharmacokinetic models. Thus, we believe that continuing to delay IRIS assessments to develop and test enhanced methodologies and strategies for communicating them conflicts with EPA's stated goal of seeking balance among fairness, accuracy, and efficient action. Specifically, it is unclear how EPA's ongoing, nearly 2-year delay of the assessment of tetrachloroethylene (perc)--a widespread groundwater contaminant-- pending the development of an acceptable uncertainty analysis represents an appropriate balance between accuracy and efficient action. 7. The draft and final reports discuss the changes EPA has made to its IRIS assessment process, including the one highlighted in EPA's letter- -a change from letter peer reviews to panel peer reviews open to the public. 8. EPA states that the current IRIS process strikes a balance between transparency and having full and frank discussions by encouraging open discussion of science and science policy questions while appropriately protecting the deliberative process. We disagree that such a balance exists because the OMB/interagency review process is not transparent at all. In its comments, EPA dismisses any impact of the opaque OMB/ interagency review process, emphasizing that all IRIS assessments undergo public and external peer review. However, the presence of transparency at a later stage of IRIS assessment development does not excuse or explain its absence earlier. Further, the National Academies have stated that the dialogue between risk assessors and risk managers should not bias or otherwise color the risk assessment conducted, and risk assessment and risk management activities should remain distinct. Transparency in the IRIS assessment process can provide assurance that these scientific assessments are appropriately based on the best available science and that they are not impacted by policy issues and considerations. Finally, because federal agencies' comments on IRIS assessments provided via the OMB/interagency process are to be scientific in nature, it is unclear why agencies would not candidly comment on an IRIS assessment even if their comments would become part of the public record. 9. EPA's comments note that transparency is applicable to both the assessment process and the content of IRIS assessments (e.g., the rationale for relying on or using specific data sets, assumptions, and models). We have revised our report to recognize that transparency is relevant to both the assessment process and the content of IRIS assessments and to be clear that the transparency issues we discuss primarily relate to the process. We note that many of the process changes EPA has made beginning in the 1990s are aimed at improving the transparency of the content of IRIS assessments as well as the process. 10. This information on EPA process improvements is provided in the draft and final reports. 11. EPA stated that final IRIS documents are now published on the IRIS public Web site and that the entire record is available to the public. This statement is not accurate because the record of the OMB/ interagency review comments and EPA's responses to them are not made available to the public. 12. The draft and final reports provide information on EPA's nomination process for new assessments, which includes a Federal Register notice. [End of section] Appendix V: Comments from the Office of Management and Budget: Executive Office Of The President: Office Of Management And Budget: Washington, D.C. 20503: February 21, 2008: Mr. John B. Stephenson: Director, Natural Resources and Environment: U.S. Government Accountability Office: 441 G Street, N.W., Room 2075: Washington, D.C. 20548: Dear Mr. Stephenson: Thank you for the opportunity to comment on the Government Accountability Office's (GAO) draft report titled, "Chemical Assessments: Low Productivity and Lack of Transparency Limit the Usefulness and Credibility of EPA's Integrated Risk Information System" (GAO-08-440). We applaud your efforts to evaluate the EPA Integrated Risk Information System (IRIS) as we believe this is a very important EPA program. As your report states, funding for the IRIS program has increased-from $1.7 million in FY 2000 to $9.6 million in FY 2007; OMB has supported changes that will improve the quality and efficiency of the IRIS program. However, the draft GAO report fundamentally misunderstands the purpose of interagency review and, unfortunately, could leave readers with a false impression of the role of OMB/interagency review in the IRIS process. We discuss below our most important concerns with the draft report. [Footnote 47] [See comment 2] First, the draft report mischaracterizes OMB's role by alleging that OMB has imposed an OMB-managed interagency review process of IRIS risk assessment documents that has resulted in a loss of EPA control. [See comment 1][Footnote 48] In fact, however, only EPA has the authority to finalize an EPA assessment and only EPA has authority to determine when a draft may be sent for external peer review. The interagency review process is a dialogue that helps to ensure the quality (including objectivity and transparency) of agency documents. Input from scientists throughout the Federal government (including, for example, HHS, DOD, DOE, NASA, DOL, and USDA within the context of the IRIS process) helps inform and improve the quality of IRIS assessments. These reviews address the objectivity, clarity, and transparency of the work, identifying key science policy issues for EPA's consideration, and scientific issues for external peer review. [Footnote 49] Simply put, the interagency review process works as follows: OMB oversees an interagency process that is part of a broader EPA process. During the interagency process, EPA works collaboratively with OMB and other agencies to achieve resolution and completion of interagency reviews. These discussions have sometimes required resolution by the EPA Science Advisor, who is also the EPA Assistant Administrator of the Office of Research and Development. When this review process is complete, OMB concludes interagency review. At that time, EPA moves to the next step - generally, release of the risk assessment for public comment and an external peer review by a group of independent experts. [Footnote 50] To be clear, EPA is entirely responsible for the content of information on IRIS. Second, the draft report alleges that the fact that OMB manages and directs the review process has "made it more difficult for EPA to complete IRIS assessments in a timely manner." [Footnote 51] [See comment 3] The draft report asserts that the OMB/interagency review adds significant time to IRIS time frames. It is our experience, however, that the interagency commenters have generally provided comments in a very timely manner. We are not aware of delays over "nonsubstantive issues" as alleged in the draft report. [Footnote 52] We also note that the draft report does not provide specific examples on which we can comment. [Footnote 53] [See comment 6] In addition, we believe the draft report is seriously flawed in that it does not consider whether interagency review operates to improve the quality of IRIS assessments, both by addressing the objectivity, clarity, and transparency of the work as well as identifying key science policy issues for EPA consideration and technical issues for peer review. [Footnote 54] [See comment 4] The draft report's focus on timeliness, without consideration of quality of the end product, presents a misleading and incomplete picture. Third, the draft report alleges that "the OMB/interagency review process also affects the credibility of assessments primarily because the review process lacks transparency." [Footnote 55] [See comment 5] To address this supposed issue, the draft report recommends that EPA make public comments by OMB and other Federal agencies to decision makers, the Congress, and the public. However, the draft report fails to acknowledge that, in the case of documents that involve interagency deliberations, these documents are covered by the deliberative process privilege, a well-recognized privilege that has been affirmed by the Congress in the Freedom of Information Act (5 U.S.C. 552(b)(5)) and by the Supreme Court in such cases as NLRB v. Sears, Roebuck & Co., 421 U.S. 132 (1975), and EPA v. Mink, 410 U.S. 73 (1973). [Footnote 56] [See comment 8] Accordingly, protection of internal Executive Branch communications is not "inconsistent with the principle of sound science," contrary to the erroneous conclusions of the draft report. [Footnote 57] In fact, the National Academy of Sciences, which is recognized as a leader in scientific reviews, uses a multi-step review process that includes deliberations that are withheld from the public. [Footnote 58] [See comment 9] Finally, the draft report discusses at some length a draft document that it alleges serves to establish a "formal IRIS assessment process" that EPA has been developing over the past two years. [Footnote 59] [See comment 7] The discussion in the draft GAO report is misleading to readers; EPA's draft document is not a final document. [Footnote 60] [See comment 10] EPA has not completed the development of this process. Indeed, the process will not be complete until EPA circulates its draft to the public for comments and then releases a final product that is responsive to those comments. Thank you again for the opportunity to review and comment on your draft report. We hope you take these comments into account in preparing your final report. Sincerely, Signed by: Kevin F. Neyland: Deputy Administrator: Office of Information and Regulatory Affairs: The following are GAO's comments on the Office of Management and Budget's letter dated February 21, 2008. GAO Comments: 1. OMB states that the draft report mischaracterizes OMB's role by alleging that OMB has imposed an OMB-managed interagency review process of IRIS risk assessment documents that has resulted in a loss of EPA control. We believe that we have appropriately described the OMB/ interagency review process and identified the time frame and assessment content limitations it has placed on EPA. First, as discussed in the draft and final reports, OMB has limited EPA's ability to determine the types of assessments it will conduct to meet its program needs by requiring EPA to terminate five IRIS assessments. Second, while EPA's annual performance goal for IRIS assessments had been the number of completed assessments, during the Program Assessment Rating Tool review by OMB it was agreed that the number of assessments provided to OMB for OMB/interagency review was the most appropriate annual measure of performance because EPA "relinquishes direct control of production dates" when it sends draft IRIS assessments to OMB. Third, our draft and final reports are clear that EPA officially determines when drafts are sent for external peer review or are finalized, but they are also clear that prior to taking these steps, EPA must be notified by OMB that EPA has adequately addressed interagency comments. Thus, the current process elevates the goal of reaching interagency agreement above achieving IRIS program objectives and in doing so fails to adequately acknowledge the expertise of EPA scientists and the many other scientific experts who have prepared and reviewed the assessments. Importantly, the fact that the first assessment drafts that become publicly available--those provided to independent peer reviewers--incorporate changes from federal agencies that may be affected by the assessments without identifying these changes or the reasons they were made weakens the credibility of the assessments. Finally, because the negotiations over OMB/interagency comments are not disclosed, OMB's assertion that EPA is entirely responsible for the content of information on IRIS is open to question. 2. GAO provides statements of facts to agencies to confirm the factual accuracy of statements upon which reports are based. OMB commented that we did not incorporate its concerns about the draft statement of facts we provided to OMB. We considered OMB's comments that addressed factual information and, in response, made some revisions as appropriate. However, many of OMB's comments did not address facts but were focused on broader issues, such as the scope and methodology of our review. We disagreed with OMB's broad comments and thus did not make certain changes OMB suggested. For example, OMB expressed concern that our report discusses the development of proposed revisions to EPA's IRIS assessment process, which has not been released for public comment. As we discuss further in comment 10, GAO audits and reviews include those providing prospective analyses of possible or planned agency actions. 3. OMB takes issue with our conclusion that the OMB/interagency review has made it more difficult for EPA to complete IRIS assessments in a timely manner. Our draft and final reports show that the initial OMB/ interagency reviews that had been completed as of December 1, 2007, added 6 months or more to the IRIS assessment process. The average length of time for these initial reviews was 7-1/2 months. Given the impact of delays on IRIS assessments and the staff time needed to respond to OMB/interagency comments, we believe such delays are significant. Further, our report discusses five key factors that make it difficult for EPA to complete assessments in a timely manner--one of which is the OMB/interagency review process. 4. OMB states its belief that the draft report is seriously flawed because it does not consider whether interagency reviews improve the quality of IRIS assessments. As OMB notes, our draft report states that we did not evaluate the scientific content or quality of IRIS assessments. However, as the draft and final reports state, the Board of Scientific Counselors--an independent board of experts appointed by EPA to assess its programs--reviewed the effectiveness of the IRIS program (as part of its review of EPA's human health risk assessment program) in November 2007. A report is expected to be finalized in early 2008. In response to our congressional request and to avoid duplication of effort, we focused on the timeliness and credibility of IRIS assessments in the context of the current assessment process and proposed revisions to it. Our draft report also noted that it is too early to determine whether or to what extent the IRIS assessment process changes EPA has implemented in recent years have enhanced the scientific credibility of the assessments--for example, only four assessments were finalized in fiscal years 2006 and 2007. Importantly, because OMB/interagency comments and EPA's response to them are not disclosed, the extent to which the comments added value or caused EPA to revise its risk estimates cannot be determined. However, the status of the IRIS assessment of dibutyl phthalate raises questions about the extent to which the OMB/interagency reviews improve the quality of IRIS assessments. Specifically, this key assessment had cleared both OMB/ interagency reviews and was in the process of being finalized in July 2007. However, because of methodology concerns raised by the Natural Resources Defense Council and the Chairman, Senate Committee on Environment and Public Works, EPA suspended the assessment pending a review by the National Academies. 5. OMB asserts that "the draft report fails to acknowledge that, in the case of documents that involve interagency deliberations, these documents are covered by the deliberative process privilege," and that "accordingly, protection of internal Executive Branch communications is not 'inconsistent with the principle of sound science.'" Contrary to OMB's assertion, the report specifically acknowledges that OMB considers the documents at issue to be protected from disclosure because of their deliberative nature. Moreover, OMB's assertions concerning the deliberative process privilege are misleading and illogical. That is, OMB's comments fail to note that the deliberative process privilege protects internal and interagency communications from judicially compelled disclosure, an issue irrelevant to our report.[Footnote 65] The privilege in no way prevents agencies from voluntarily disclosing such information.[Footnote 66] OMB is thus arguing that because the scientific comments at issue might generally be protected from discovery in civil litigation, refusal to disclose them voluntarily in this specific context is necessarily consistent with the principles of sound science. OMB provides no citation or other support for this conflation of judicial and scientific procedures. Moreover, OMB's comments that the deliberative process privilege was affirmed in the Freedom of Information Act (FOIA) fails to acknowledge that FOIA requires federal agencies to disclose factual material in documents covered by the deliberative process privilege if the factual material can be reasonably segregated from the deliberative material.[Footnote 67] OMB's comment appears to assume that because a document passes from one agency to another all of its contents may be withheld from disclosure, a legally unsupportable assertion that courts have uniformly rejected, particularly in the context of attempts to withhold scientific information.[Footnote 68] Indeed, to the extent that interagency comments on IRIS assessments reflect political and policy concerns rather than discussions of scientific issues, they would be easier to withhold from disclosure in court but harder to justify as part of the IRIS scientific assessment process. In addition, OMB's comments do not distinguish between risk assessment and risk management decisions, such as the formulation of regulations. As discussed in the draft and final reports, IRIS assessments are not themselves regulations, and OMB and other agencies will have opportunities to engage in deliberative policy debates during, for example, interagency reviews that occur during rule makings. These are points in the process where it is appropriate to address policy questions, such as the implications for other agencies of a specific regulatory decision. Finally, under the interagency process as currently contemplated, some IRIS assessment reviewers--representatives of federal agencies-- essentially are given favored status. OMB fails to explain why certain scientific views should be given added consideration and protected from the critical scientific scrutiny all other comments will receive simply because the reviewers providing the comments are federal employees. 6. OMB stated general disagreement with information presented in our discussion of the naphthalene assessment but did not cite the specific information with which it disagreed. Instead, OMB referred readers to the charge and report from the second peer review EPA had conducted on this chemical at OMB's direction. In our draft and final reports, we report EPA officials' description of the purpose and conclusions of this peer review. Specifically, EPA officials said (1) the agency convened an additional expert review panel on genotoxicity to obtain recommendations about short-term tests that OMB thought could be done quickly and (2) the panel concluded that such research could not be conducted in the short term. The questions posed to the peer reviewers and the summary of results in the report cited by OMB are consistent with EPA's description. 7. OMB says that the draft report is misleading to readers because EPA's "draft document is not a final document." The report clearly identifies the proposed IRIS assessment changes that EPA has been working on. For example, the graphic outlining the proposed changes is titled "EPA's Draft Proposed IRIS Assessment Process." The draft and final reports also state that "as of December 2007, EPA was planning to send a revised draft back to the interagency work group for review and approval." However, to ensure clarity, we have added "draft" prior to our use of the term "proposed process" in a number of places. 8. We did not assert that the OMB/interagency process is equivalent to peer review, and we agree with OMB that this review process happens before any external peer review occurs. However, because the OMB/ interagency process is opaque, neither peer reviewers nor the public are privy to the changes EPA made to the draft assessments or the charge questions to the peer review panels in response to the comments from OMB and other federal agencies. The presence of transparency at a later stage of IRIS assessment development does not excuse or explain its absence earlier. 9. It is unclear why OMB attempts to rely on the Federal Advisory Committee Act Amendments of 1997 to defend the lack of transparency in the interagency IRIS review process. The transparency procedures applicable to the National Academies' committees to which OMB refers far exceed those that exist under the OMB/interagency IRIS process. For example, under the legislation OMB cites, National Academies' committees must (1) provide the names, biographies, and conflict of interest disclosures of committee appointees; (2) provide an opportunity for the public to comment on the proposed committee member appointments; (3) ensure that meetings focused on data gathering are generally open to the public; (4) provide the names of reviewers of draft committee reports; and (5) provide summaries of any closed committee meetings. 5 U.S.C. App. 2, § 15. The interagency portion of the IRIS process does none of these things. Moreover, while the act authorizes a National Academies' committee to close meetings at which information exempt from disclosure under FOIA would be discussed, the President of the National Academy of Sciences assured the bill's sponsor, in a letter commenting on the House bill that would later be enacted into law, that the Academy would not rely on the deliberative process exemption as the basis for closing a meeting. "I wish to assure you that we subscribe fully to the goal of providing as much openness as possible in our work. In particular, we have no intention of using Section 552(b)(5), which deals with interagency memoranda, as a basis for closing meetings of Academy committees. In fact, it is the Academy's standard practice not to treat the type of material covered by Section 552(b)(5) as confidential input to any Academy deliberative process. This procedure insures that, inasmuch as possible, all the information that a committee uses to reach its conclusion is in the public record." 10. We were asked by the Chairman, Senate Committee on Environment and Public Works, to examine the potential effects of planned changes to the IRIS assessment process on EPA's ability to ensure that IRIS provides current, credible risk information. This review therefore involved evaluating a draft of EPA's planned IRIS assessment changes. GAO audits and reviews include those providing prospective analyses of possible or planned agency actions. We may also assess the ability of alternative approaches to yield better program performance or eliminate factors that inhibit program effectiveness. As is typical in cases in which we evaluate draft proposals, we make our recommendations on the IRIS assessment process in the spirit of informing those revisions. [End of section] Appendix VI: GAO Contact and Staff Acknowledgments: GAO Contact: John B. Stephenson, (202) 512-3841 or stephensonj@gao.gov: Staff Acknowledgments: In addition to the contact named above, Christine Fishkin (Assistant Director), Mark Braza, Nancy Crothers, Laura Gatz, Richard P. Johnson, Summer Lingard, Susan Swearingen, and Delia Zee made key contributions to this report. Also contributing to this report were Tim Bober, Phylis Cline, Michael Derr, and Cynthia Taylor. [End of section] Footnotes: [1] The National Academies comprises four organizations: the National Academy of Sciences, the National Academy of Engineering, the Institute of Medicine, and the National Research Council. [2] Transparency is relevant to both the IRIS assessment process (for example, the public availability of information about the assessment process, the input from external reviews on draft assessments, and EPA's responses to them) and the content of IRIS assessments (for example, the rationale for using specific data sets, assumptions, or models). In this report, the transparency issues we discuss primarily relate to the IRIS assessment process. [3] The cancer slope factor and unit risk typically are both upper bound estimates (the plausible statistical upper limits of the true value of a quantity). [4] OMB, Fiscal Year 2006 Program Assessment Rating Tool (PART) assessment of EPA's Human Health Risk Assessment Program. [5] This internal peer review continues to be conducted during the assessment drafting stage. [6] OMB did not include these five assessments towards EPA's fiscal year 2006 annual performance goals and thus does not acknowledge that EPA sent 16 draft assessments to it for OMB/interagency review that fiscal year. [7] EPA has, in some cases, divided an ongoing assessment into two assessments. For example, in fiscal year 2007, the arsenic assessment covering both cancer and noncancer risks, which was started in 2003, was divided into two separate assessments. Although the noncancer assessment was started in 2003, IRIS Track, which tracks the status of ongoing assessments, shows a 2007 start date for the IRIS noncancer assessment. [8] The two offices identified the same priority chemicals in eight cases. [9] The other nominations were for chemicals in IRIS with outdated assessments. [10] The preliminary findings of the Use of IRIS Project were presented during the EPA's Board of Scientific Counselors' Human Health Risk Assessment Subcommittee meeting on November 15, 2007. [11] EPA, Human Health Risk Assessment Multi-Year Plan (Washington, D.C., 2007). [12] In its 2005 multiyear plan EPA, specified that it would send the assessments for acrylamide, MTBE, naphthalene, tetrachloroethylene (perc), acrylonitrile, formaldehyde, methanol, methylene chloride, trichloroethylene, and dioxin for external review during fiscal years 2006 and 2007. Only acrylamide and tetrachloroethylene (perc) have been sent to OMB for OMB/interagency review. [13] Under the executive order, significant guidance is defined as, "a guidance document disseminated to regulated entities or the general public that may reasonably be anticipated to (1) lead to an annual effect of $100 million or more or adversely affect in a material way the economy, a sector of the economy, productivity, competition, jobs, the environment, public health or safety, or State, local, or tribal governments or communities; (2) create a serious inconsistency or otherwise interfere with an action taken or planned by another agency; (3) materially alter the budgetary impacts of entitlements, grants, user fees or loan programs or the rights or obligations of recipients thereof; or (4) raise novel legal or policy issues arising out of legal mandates, the President's priorities, or the principles set forth in this Executive order." [14] Examples cited by the former director include scientific data such as subtle biologic changes, biomarkers, and partial data on one or more mechanistic hypotheses. [15] Examples of PBPK model applications in risk assessments include interspecies extrapolation of the dose-response relationship, route- to- route extrapolation, estimation of response from varying exposure conditions, estimation of human variability (within the whole population or subpopulations), and high-to-low dose extrapolation. [16] EPA plans to release draft reports in 2010 on methods for analyzing and characterizing uncertainty in hazard and dose response and in PBPK models. [17] According to an EPA official, the November 2007 review of the IRIS program by the Board of Scientific Counselors, an independent board of experts appointed by EPA to assess its program, concluded that the program "met expectations." A report covering this review, which included evaluating the scientific quality of IRIS assessments, is expected in early 2008. [18] Also referred to as RDX or hexahydro-1,3,5-trinitrotriazine. [19] The Board of Scientific Counselors is a federal advisory committee established by EPA to provide advice, information, and recommendations about the Office of Research and Development's research program. [20] OMB recently proposed a bulletin that would have provided risk assessment guidance to federal agencies. The proposed bulletin stated that "every quantitative risk assessment should provide a range of plausible risk estimates when there is scientific uncertainty or variability." OMB decided not to finalize the bulletin after a National Academies' committee severely criticized it, noting, among other shortcomings, that "the description of uncertainty and variability in the bulletin is oversimplified and does not recognize the complexities of different types of risk assessments or the need to tailor uncertainty analysis to a given agency's particular needs." In particular, the committee noted that "a central estimate and a risk range might be misleading in situations when sensitive populations are of primary concern." National Academies, Scientific Review of the Proposed Risk Assessment Bulletin from the Office of Management and Budget Committee to Review the OMB Risk Assessment Bulletin (Washington, D.C., 2007). [21] In January 2006, the Assistant Administrator for Research and Development was appointed as EPA's Science Advisor. [22] EPA officials said that in July 2007, approval authority to send draft assessments to OMB was again devolved to the Director, NCEA. [23] According to the fiscal year 2004, 2005, and 2006 standard operating procedures for IRIS assessments, before an assessment is posted on the Web site, chemical managers are required to (1) brief the IRIS Program Director and other NCEA management on any significant changes since external review and (2) prepare a fact sheet for review and approval by the Assistant Administrator, Office of Research and Development, before the IRIS Program Director submits the final assessment documents to the IRIS Web master for posting on the IRIS Web site. [24] A December 1998 notice stated that EPA planned to complete the chloroform assessment, as well as 23 other assessments, in fiscal year 1999 or fiscal year 2000. In 2001, EPA completed assessments addressing oral exposure to chloroform (oral RfD and cancer assessments); however, the assessment addressing inhalation exposure (RfC) needed by the Office of Air and Radiation is currently being drafted. Since 2002, EPA has been reporting the chloroform (inhalation route) assessment as being under way or generally complete and planned for entry into IRIS within a year or 2. [25] National Academies, Scientific Review of the Proposed Risk Assessment Bulletin from the Office of Management and Budget Committee to Review the OMB Risk Assessment Bulletin (2007). [26] The European Union and California restrictions cover five phthalates, including dibutyl phthalate. [27] IRIS assessments cover individual chemicals, not cumulative exposure to multiple related chemicals. This is an emerging risk assessment issue that will require developing new assessment methodologies. The congressional letter expressed concerns that the draft IRIS assessment would not adequately protect public health but did not request that it be entirely redone. Among other things, the Natural Resources Defense Council letter suggested that EPA revisit its decisions regarding the uncertainty factors used. [28] Presentations at the Society for Risk Analysis meting reflect the views of the authors and "do not necessarily reflect the views of any other organization or agency." [29] Using its 1996 Proposed Guidelines for Carcinogen Risk Assessment, EPA concluded in the 1998 IRIS assessment of naphthalene that its human carcinogenic potential could not be determined at that time, but noted that there was suggestive evidence of potential human carcinogenicity. (EPA also noted that under its 1986 cancer guidelines, EPA classified naphthalene as a possible human carcinogen.) Subsequently, in 2002, the International Agency for Research on Cancer (IARC), part of the World Health Organization, concluded that naphthalene is possibly carcinogenic to humans; in 2004, the Department of Human Health and Services' National Toxicology Program concluded that naphthalene can reasonably be anticipated to be a human carcinogen. EPA's current assessment will be subject to the agency's 2005 cancer guidelines. [30] According to DOD, EPA did not specifically ask the peer reviewers to address some of the technical questions DOD had raised and wanted the peer review to address. [31] Genotoxic substances are a type of carcinogen, specifically those capable of causing genetic mutation and of contributing to the development of tumors. This includes both certain chemical compounds and certain types of radiation. [32] The National Priorities List is EPA's list of seriously contaminated sites. [33] The cancer portion of the formaldehyde assessment was originally issued in 1989 and updated in 1991; the noncancer assessment was added in 1990. [34] NCI published the results of its study in two publications. The first study, published in November 2003, focused on the association between formaldehyde exposure and leukemia. M. Hauptmann, J. H. Lubin, P. A. Stewart, R. B. Hayes, A. Blair, "Mortality from Lymphohematopoietic Malignancies among Workers in Formaldehyde Industries," Journal of the National Cancer Institute (2003). The second study, published in June 2004, evaluated the association between formaldehyde exposure and other cancers--including nasopharyngeal cancer. M. Hauptmann, J. H. Lubin, P. A. Stewart, R. B. Hayes, A. Blair, "Mortality from Solid Cancers among Workers in Formaldehyde Industries," American Journal of Epidemiology (2004). The results of the NIOSH study were described in one publication, dated March 2004, which assessed mortality from all causes and all cancers. L. E. Pinkerton, M. J. Hein, L. T. Stayner, "Mortality among a Cohort of Garment Workers Exposed to Formaldehyde: an Update," Occupational and Environmental Medicine (2004). [35] Natural Resources Defense Council v. E.P.A., 489 F.3d 1364, 1372- 73 (D.C. Cir, 2007). The court did not specifically address EPA's reliance on the CIIT study, holding instead that the Clean Air Act prohibited establishment of the exemptions at issue. [36] Physiologically based pharmacokinetic models are a class of dosimetry models that are useful for predicting internal doses to target organs. With the appropriate data, these models can be used to extrapolate across species and exposure scenarios and address various sources of uncertainty in risk assessments. [37] GAO, Human Health Risk Assessment: EPA Has Taken Steps to Strengthen Its Process, but Improvements Needed in Planning, Data Development, and Training, GAO-06-595 (Washington, D.C.: May 31, 2006). [38] Developing this process, publishing a notice about it in the Federal Register, and holding a public meeting to obtain public input on the proposed process has been an item OMB has included in its PART review of EPA's human health risk assessment program. [39] Under the proposal, EPA could also identify chemicals "of major importance" which would be treated as mission-critical chemicals. [40] Under EPA's draft proposed IRIS assessment process, EPA would develop, for all chemicals, an additional draft IRIS assessment--one that includes only qualitative information--that must undergo internal review. Only draft qualitative assessments for chemicals that are identified as mission critical would be released for the error correction step. [41] According to DOD, the Materials of Emerging Regulatory Interest Team is a virtual interagency team composed of individuals throughout DOD with a common interest in emerging contaminants. [42] Other possible types of DOD risk management investments include material and process substitutions and personal protective equipment. [43] The National Academies recently addressed this issue as follows: "The committee believes that risk assessors and risk managers should talk with each other; that is, a 'conceptual distinction' does not mean establishing a wall between risk assessors and risk managers. Indeed they should have constant interaction. However, the dialogue should not bias or otherwise color the risk assessment conducted, and the activities should remain distinct; that is, risk assessors should not be performing risk management activities." National Academies, Scientific Review of the Proposed Risk Assessment Bulletin from the Office of Management and Budget Committee to Review the OMB Risk Assessment Bulletin (2007). [44] The National Academies recently emphasized that "stakeholders from all points on the spectrum of interested parties--other state and federal agencies, advocacy groups from industry, and affected communities--can be expected to offer perspectives on the risk assessment policies under discussion." National Academies, Scientific Review of the Proposed Risk Assessment Bulletin. [45] "Other information" includes economic information on the costs and benefits of mitigating the risk, technological information on the feasibility of managing the risk, and the concerns of various stakeholders. [46] These five assessments address short-term health risks. According to EPA, while OMB had previously agreed that these assessments would be included in EPA's 2006 annual performance goals that OMB uses to evaluate EPA's performance, in November 2007, OMB told EPA that it would not count EPA's short-term assessments toward meeting its goals, thereby lowering EPA's performance rating by OMB. [47] In January 2008, GAO provided OMB staff with the opportunity to provide comments on the draft report titled "Statement of Facts for GAO's Review of EPA's Integrated Risk Information System." As you were aware, from our discussion on January 18, 2008, OMB staff had a number of concerns about the draft statement of facts, which were discussed in the January 22, 2008 comments that were sent to you. It appears that many of those comments have not been incorporated into the current draft report and our concerns remain. But rather than reiterating them all here, 1 will highlight our overarching comments on the draft report. [48] Draft report, pp. 22-27. [49] OMB has a broad interest in the quality, objectivity, utility and integrity of information disseminated by Federal Agencies. See OMB's Guidelines for Ensuring and Maximizing the Quality, Objectivity, Utility, and Integrity of Information Disseminated by Federal Agencies (2002), 67 Fed. Reg. 8452 (Feb. 22, 2002), available at [hyperlink, http://www.whitehouse.gov/omb/fedreg/reproducible2.pdf]; OMB's Final Information Quality Bulletin for Peer Review (2002), 70 Fed. Reg. 2664 (Jan. 14, 2005), available at [hyperlink, http://www.whitehouse.goviomb/fedreg/2005/011405_peer.pdf]; and the OMB/OSTP Updated Principles for Risk Analysis, available at [hyperlink, http://www.whitehouse.gov/omb/memoranda/fy2007/m07-24.pdf]. [50] OMB recognizes that there is also a second round of interagency review in the current IRIS process. This review seeks to ensure that comments from expert reviewers and the public have been transparently and objectively addressed. [51] Draft report, p. 21. [52] Draft report, p. 25. [53] In fact, when the draft report does provide a specific example, such as its discussion of naphthalene, it does so inaccurately. Draft report, pp. 36-38. OMB staff does not agree with the characterizations as presented in the draft report and suggests that readers would benefit from looking at the EPA documentation.: (i) the EPA charge for the peer consultation, which resulted from interagency dialogue, available at [hyperlink, http://oaspub.epa.gov/eims/eimscomm.getfile?p_download_id=46877]; and (ii) the peer consultation report, available at [hyperlink, http://oaspub.epa.gov/eims/eimscomm.getfile?p_download_id=442588]. [54] In fact, the draft report recognizes that it "did not evaluate the scientific content or quality of IRIS assessments." Draft report, p. 3. [55] Draft report, p. 26. [56] We do not consider the interagency process to constitute peer review; rather the document that EPA releases at a later step in the process will be subject to external peer review and public comment. EPA has a very transparent process by which draft IRIS assessments are released for public comment and external peer review before they are issued as final documents. The credibility of IRIS assessment is in part based on this rigorous and transparent external peer review and public comment process. [57] Draft report page 6. [58] See [hyperlink, http://www.iom.edu/CMS/6008.aspx] and [hyperlink, http://www.nationalacademies.org/studvcommittenrocess.pdf]. In 1997, Congress ensured that NAS deliberations have heightened confidentiality protections, when Congress passed stand-alone legislation that added a new section (Section 15) to the Federal Advisory Committee Act (5 U.S.C. App.) which specifically, and solely, addresses NAS and the National Academy of Public Administration. See the Federal Advisory Committee Act Amendments of 1997, Public Law 105-153, 111 Stat. 2689. The legislative history explains that Congress granted this protection "to preserve the quality of the research provided to the Federal Government through the National Academy of Sciences and the National Academy of Public Administration." Statement of Rep. Horn, Cong. Rec. 1110579 (daily ed. November 9, 1997). Representative Horn explained that "[t]he administration, the House and the Senate, both the majority and minority, all agree the academy should not be subject to the full process of the Federal Advisory Committee Act." Id. As Rep. Horn also noted, id., Congress passed this legislation in response to the court decision earlier that year in Animal Legal Defense Fund, Inc. v. Shalala, 104 F3d 1209 (D.C. Cir.), cert. denied, 522 U.S. 949 (1997), in which the court held that NAS panels were subject to the full range of FACA disclosure requirements. In its decision, the court stated that NAS, in arguing for why it should not be subject to these requirements, referred to "the deleterious effects of FACA's requirements on its deliberative processes: As the NAS sees it, open meetings and records would compromise its internal review procedures and inhibit candid exchange among its members." Id. at 428. [59] Draft report, pp. 45-52. [60] By way of comparison, we note that, in providing to us the draft report for our comment, your cover page stated that the draft "is restricted to official use only"; is "subject to revision"; and "must be safeguarded to prevent improper disclosure." [61] NLRB v. Sears, Roebuck & Co., 421 U.S. 132, 149 ("it is reasonable to construe Exemption 5 to exempt those documents, and only those documents, normally privileged in the civil discovery context"); EPA v. Mink, 410 U.S. 73, 86 (1973) (The Freedom of Information Act (FOIA) contemplates that the public is entitled to all memoranda or letters that a private party could discover in litigation with the agency). [62] See Chrysler Corp. V. Brown, 441 U.S. 281, 293 (holding that exemptions from FOIA's general requirement to disclose agency documents do not prohibit agencies from disclosing information covered by the exemptions). In fact, OMB has voluntarily released information it considered to be protected from disclosure under FOIA, presumably because it saw some benefit in doing so. NRDC v. U.S. Department of Defense, 442 F.Supp.2d 857, 863 (C.D. Cal. 2006). It is unclear from the case which FOIA exemptions OMB believed were applicable. [63] E.g., Mead Data Central, Inc. v. U.S. Dept. of Air Force, 566 F.2d 242, 260 (D.C. Cir. 1977) ("The focus of the FOIA is information, not documents, and an agency cannot justify withholding an entire document simply by showing that it contains some exempt material."). [64] E.g., Bristol-Myers v. FTC, 424 F.2d 935, 939 (FOIA "does not authorize an agency to throw a protective blanket over all information by casting it in the form of an internal memorandum. Purely factual reports and scientific studies cannot be cloaked in secrecy by an exemption designed to protect only those internal working papers in which opinions are expressed and policies formulated and recommended"); Southwest Center v. Biological Diversity v. USDA, 170 F.Supp.2d 931, 943 (D. Ariz. 2000) ("FOIA exemption five does not protect research data"); Verrazzano Trading Corp. v. United States, 349 F.Supp. 1401, 1406 (Cust. Ct. 1972) ("the exemption was not intended to protect factual or scientific reports and investigations"). [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office: 441 G Street NW, Room LM: Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548: