Military Readiness: Congress Needs Better Tools for Effective Oversight (Testimony, 03/18/98, GAO/T-NSIAD-98-124). GAO discussed the Department of Defense's (DOD) process for assessing and reporting on military readiness, focusing on: (1) what corrective action DOD has taken to improve its readiness assessment system; (2) whether military readiness reports provided quarterly to Congress effectively support congressional oversight; and (3) whether further improvements are needed to DOD's process. GAO noted that: (1) over the last few years, DOD has taken action to improve readiness assessment; (2) DOD has made technical enhancements to the Status of Resources and Training System (SORTS)--the automated system it uses to assess readiness at the unit level; (3) DOD also has established two forums--the Joint Monthly Readiness Review and the Senior Readiness Oversight Council--for evaluating readiness from a joint and strategic perspective; (4) however, SORTS remains the basic building block for readiness assessment, and inherent limitations to this system, such as its inability to signal impending changes in readiness and its imprecise ratings for unit resources and training, may be reflected in reviews at the joint and strategic levels; (5) DOD's quarterly reports to Congress, which are based on information provided to the Senior Readiness Oversight Council, provide only a vague description of readiness deficiencies and planned remedial actions; consequently, in their present form they are not as effective as they could be as a congressional oversight tool; (6) DOD is required to expand on these reports beginning in October 1998 by adding indicators mandated by Congress; (7) GAO has concerns about DOD's current plans for implementing this expanded reporting requirement; (8) for example, current plans do not present a clear picture of how the additional readiness will be incorporated into the quarterly report; (9) GAO's work has identified two areas in which DOD can improve its readiness reporting to Congress; (10) DOD should provide more specific descriptions and supporting information for the key readiness deficiencies and planned remedial actions identified in its quarterly report; and (11) DOD can make improvements to its current plans for adding readiness indicators to the quarterly report. --------------------------- Indexing Terms ----------------------------- REPORTNUM: T-NSIAD-98-124 TITLE: Military Readiness: Congress Needs Better Tools for Effective Oversight DATE: 03/18/98 SUBJECT: Evaluation methods Combat readiness Defense contingency planning Personnel evaluation systems Management information systems Reporting requirements Defense capabilities Military forces Training utilization Military training IDENTIFIER: DOD Status of Resources and Training System ****************************************************************** ** This file contains an ASCII representation of the text of a ** ** GAO report. Delineations within the text indicating chapter ** ** titles, headings, and bullets are preserved. Major ** ** divisions and subdivisions of the text, such as Chapters, ** ** Sections, and Appendixes, are identified by double and ** ** single lines. The numbers on the right end of these lines ** ** indicate the position of each of the subsections in the ** ** document outline. These numbers do NOT correspond with the ** ** page numbers of the printed product. ** ** ** ** No attempt has been made to display graphic images, although ** ** figure captions are reproduced. Tables are included, but ** ** may not resemble those in the printed version. ** ** ** ** Please see the PDF (Portable Document Format) file, when ** ** available, for a complete electronic file of the printed ** ** document's contents. ** ** ** ** A printed copy of this report may be obtained from the GAO ** ** Document Distribution Center. For further details, please ** ** send an e-mail message to: ** ** ** **** ** ** ** with the message 'info' in the body. ** ****************************************************************** Cover ================================================================ COVER Before the Subcommittee on Military Readiness, Committee on National Security, House of Representatives For Release on Delivery Expected at 3 p.m., EDT Wednesday, March 18, 1998 MILITARY READINESS - CONGRESS NEEDS BETTER TOOLS FOR EFFECTIVE OVERSIGHT Statement of Mark E. Gebicke, Director, Military Operations and Capabilities Issues, National Security and International Affairs Division GAO/T-NSIAD-98-124 GAO/NSIAD-98-124T Military Readiness (703234) Abbreviations =============================================================== ABBREV DOD - Department of Defense LMI - Logistics Management Institute GSORTS - Global Status of Resources and Training System OPTEMPO - operations tempo SORTS - Status of Resources and Training System ============================================================ Chapter 0 Mr. Chairman and Members of the Subcommittee: I am pleased to be here today to discuss the Department of Defense's (DOD) process for assessing and reporting on military readiness. Questions about the validity and thoroughness of this process have been raised for some time now. Various audit and oversight organizations, including our office, have periodically reported on limitations to official unit readiness reports. Congress has expressed concern regarding apparent inconsistencies between official readiness reports and the actual readiness of units in the field. In addition, Congress has expressed concern about DOD's lack of progress in integrating additional readiness indicators into official readiness reports and recently passed legislation requiring DOD to include the indicators in its reports to Congress. With this history in mind, I would like to frame my comments today around three questions: -- What corrective actions has DOD taken to improve its readiness assessment system? -- Do military readiness reports provided quarterly to Congress effectively support congressional oversight? -- Are further improvements to DOD's reporting process needed? Before I answer these questions, some background on DOD's readiness assessment system may be helpful. DOD assesses military readiness at three levels--(1) the individual unit level; (2) the joint force level; and (3) the aggregate, or strategic, level. "Unit readiness" refers to the ability of units, such as Army divisions, Navy ships, and Air Force wings, to provide capabilities required of the combatant commands and is derived from the ability of each unit to deliver the outputs for which it was designed. "Joint readiness" is the combatant commands' ability to integrate and synchronize units from one or more services to execute missions. "Strategic readiness" is a synthesis of unit and joint readiness and concerns the ability of the armed forces as a whole, including the services, the combatant commands, and the combat support agencies, to fight and meet the demands of the national security strategy. SUMMARY ---------------------------------------------------------- Chapter 0:1 Over the last few years, DOD has taken action to improve readiness assessment. DOD has made technical enhancements to the Status of Resources and Training System (SORTS)--the automated system it uses to assess readiness at the unit level.\1 DOD also has established two forums--the Joint Monthly Readiness Review and the Senior Readiness Oversight Council--for evaluating readiness from a joint and strategic perspective. We believe these changes represent progress. However, SORTS remains the basic building block for readiness assessment, and inherent limitations to this system, such as its inability to signal impending changes in readiness and its imprecise ratings for unit resources and training, may be reflected in reviews at the joint and strategic levels. DOD's quarterly reports to Congress, which are based on information provided to the Senior Readiness Oversight Council, provide only a vague description of readiness deficiencies and planned remedial actions; consequently, in their present form they are not as effective as they could be as a congressional oversight tool. DOD is required to expand on these reports beginning October 1998 by adding indicators mandated by Congress. We have concerns about DOD's current plans for implementing this expanded reporting requirement. For example, current plans do not present a clear picture of how the additional readiness indicators will be incorporated into the quarterly report. Our work has identified two areas in which DOD can improve its readiness reporting to Congress. First, DOD should provide more specific descriptions and supporting information for the key readiness deficiencies and planned remedial actions identified in its quarterly report. Second, DOD can make improvements to its current plans for adding readiness indicators to the quarterly report. Let me address these issues in more detail. -------------------- \1 SORTS officially evolved into the Global Status of Resources and Training System (GSORTS) with the advent of the Global Command and Control System. We use the term SORTS throughout this statement because it is more familiar and continues to be commonly used. WHAT CORRECTIVE ACTIONS HAS DOD TAKEN TO IMPROVE ITS READINESS ASSESSMENT SYSTEM? ---------------------------------------------------------- Chapter 0:2 Any discussion of readiness measurement must start with SORTS. This automated system, which functions as the central listing for more than 9,000 military units, is the foundation of DOD's unit readiness assessment process and is a primary source of information used for reviews at the joint and strategic levels. The system's database indicates, at a selected point in time, the extent to which these units possess the required resources and training to undertake their wartime missions. Units regularly report this information using a rating system that comprises various indicators on the status of personnel, equipment, supplies, and training. SORTS is intended to enable the Joint Staff, the combatant commands, and the military services to, among other things, prepare lists of readily available units, assist in identifying or confirming major constraints on the employment of units, and confirm shortfalls and distribution problems with unit resources. Until the early 1990s, DOD defined "readiness" narrowly in terms of the ability of units to accomplish the missions for which they were designed, and SORTS was the only nonservice-specific system DOD had to measure readiness. Even today, SORTS remains an important component of readiness assessment in that data from the system is used extensively by the services to formulate a big-picture view of readiness. However, limitations to SORTS have been well documented for many years by various audit and oversight organizations. For example, prior reviews by our office and others have found: -- SORTS represents a snapshot in time and does not signal impending changes in readiness. -- SORTS relies on military judgment for certain ratings, including the commanders' overall rating of unit readiness. In some cases, SORTS ratings reflect a higher or lower rating than the reported analytical measures support. However, DOD officials view subjectivity in SORTS reports as a strength because the commanders' judgments provide professional military assessments of unit readiness. The officials also note that much of the information in the SORTS reports is objective and quantitative. -- The broad measurements that comprise SORTS ratings for resource availability may mislead managers because they are imprecise and therefore may mask underlying problems. For example, SORTS allows units to report the same capability rating for personnel strength even though their personnel strength may differ by 10 percent. -- SORTS data is maintained in multiple databases located at combatant commands, major commands, and service headquarters and is not synchronized across the databases. -- SORTS data may be out-of-date or nonexistent for some units registered in the database because reporting requirements are not enforced. -- Army SORTS procedures that require review of unit reports through the chain of command significantly delay the submission of SORTS data to the Joint Staff. DOD is taking actions to address some of these limitations. The Chairman of the Joint Chiefs of Staff was directed last year--in the Defense Planning Guidance--to develop a plan for improving DOD's readiness assessment system. Although it has yet to be approved, the Joint Staff plan calls for a phased improvement to the readiness assessment system, starting with upgrades to SORTS. During the first phase of the plan, the Joint Staff is addressing technical limitations of SORTS. One of the objectives, for instance, is to ensure that the data is synchronized DOD-wide across multiple databases. Future phases of the Joint Staff plan would link SORTS with other databases in a common computer environment to make readiness information more readily accessible to decisionmakers. In addition, the Joint Staff plan calls for upgrades to SORTS that will make the system easier to use. Separate from the Joint Staff plan, the services are developing or implementing software to automate the process of entering SORTS data at the unit level. These technical upgrades are aimed at improving the timeliness and accuracy of the SORTS database and, therefore, are positive steps. They, however, will not address some of the inherent limitations to the system. For instance, the upgrades will not address the inability of the system to signal impending changes in readiness. In addition, the upgrades will not address the lack of precision in reporting unit resources and training. Another step DOD has taken to improve its readiness assessment capability is to institute a process known as the Joint Monthly Readiness Review. The joint review was initiated toward the end of 1994 and has matured over the last year or so. It represents DOD's attempt to look beyond the traditional unit perspective provided by SORTS--although SORTS data continues to play an important role--and to introduce a joint component to readiness assessment. We believe the joint review process has several notable features. First, it brings together readiness assessments from a broad range of DOD organizations and elevates readiness concerns to senior military officials, including the Vice Chairman of the Joint Chiefs of Staff. Second, the joint review emphasizes current and near-term readiness and incorporates wartime scenarios based on actual war plans and existing resources. Third, it adds a joint perspective by incorporating readiness assessments from the combatant commands. The services and combat support agencies also conduct readiness assessments for the joint review. Fourth, the joint review is conducted on a recurring cycle--four times a year--that has helped to institutionalize the process of readiness assessment within DOD.\2 Finally, the joint review includes procedures for tracking and addressing reported deficiencies. I would like to note, however, that the DOD components participating in the review are accorded flexibility in how they conduct their assessments. The 11 combatant commands, for instance, assess readiness in eight separate functional areas, such as mobility, infrastructure, and intelligence, surveillance, and reconnaissance, and to do this each command has been allowed to independently develop its own measures. In addition, the process depends heavily on the judgment of military commanders to formulate their assessment. Officials involved with the joint review view this subjectivity as a strength, not a weakness, of the process. They said readiness assessment is influenced by many factors, not all of which are readily measured by objective indicators. One consequence, however, is that the joint review cannot be used to make direct comparisons among the commands in the eight functional areas. We should also point out that the services, in conducting their portion of the joint review, depend extensively on SORTS data. As I mentioned earlier, SORTS has certain inherent limitations. -------------------- \2 The other months of the joint review cycle are devoted to working and providing feedback on reported deficiencies. DO MILITARY READINESS REPORTS PROVIDED TO CONGRESS EFFECTIVELY SUPPORT CONGRESSIONAL OVERSIGHT? ---------------------------------------------------------- Chapter 0:3 DOD is required under 10 U.S.C. 482 to prepare a quarterly readiness report to Congress. Under this law, DOD must specifically describe (1) each readiness problem and deficiency identified, (2) planned remedial actions, and (3) the key indicators and other relevant information related to each identified problem and deficiency. In mandating the report, Congress hoped to enhance its oversight of military readiness. The first report was submitted to Congress in May 1996. DOD bases its quarterly reports on briefings to the Senior Readiness Oversight Council. The Council, comprising senior civilian and military leaders, meets monthly and is chaired by the Deputy Secretary of Defense. The briefings to the Council are summaries from the Joint Monthly Readiness Review. In addition, the Deputy Secretary of Defense periodically tasks the Joint Staff and the services to brief the Council on various readiness topics. From these briefings, the Joint Staff drafts the quarterly report. It is then reviewed within DOD before it is submitted to Congress. We recently reviewed several quarterly reports to determine whether they (1) accurately reflect readiness information briefed to the Council and (2) provide information needed for congressional oversight. Because minutes of the Council's meetings are not maintained, we do not know what was actually discussed. Lacking such records, we traced information in the quarterly readiness reports to the briefing documents prepared for the Council. Our analysis showed that the quarterly reports accurately reflected information from these briefings. In fact, the quarterly reports often described the issues using the same wording contained in the briefings to the Council. The briefings, as well as the quarterly reports, presented a highly aggregated view of readiness, focusing on generalized strategic concerns. They were not intended to and did not highlight problems at the individual combatant command or unit level. DOD officials offered this as an explanation for why visits to individual units may yield impressions of readiness that are not consistent with the quarterly reports. Our review also showed that the quarterly reports did not fulfill the legislative reporting requirements under 10 U.S.C. 482 because they lacked the specific detail on deficiencies and planned remedial actions needed for congressional oversight. Lacking such detail, the quarterly reports provided Congress with only a vague picture of DOD's readiness problems. For example, one report stated that Army personnel readiness was a problem, but it did not provide data on the numbers of personnel or units involved. Further, the report did not discuss how the deficiency affected the overall readiness of the units involved. Also, the quarterly reports we reviewed did not specifically describe planned remedial actions. Rather, they discussed remedial actions only in general terms, with few specific details, and provided little insight into how DOD planned to correct the problems. Congress has taken steps recently to expand the quarterly reporting requirements in 10 U.S.C. 482. Beginning in October 1998, DOD will be required to incorporate 19 additional readiness indicators in the quarterly reports. To understand the rationale for these additional indicators, it may be helpful to review their history. In 1994, we told this Subcommittee that SORTS did not provide all the information that military officials believed was needed for a comprehensive assessment of readiness. We reported on 26 indicators that were not in SORTS but that military commanders said were important for a comprehensive assessment of readiness. We recommended that the Secretary of Defense direct his office to determine which indicators were most relevant to building a comprehensive readiness system, develop criteria to evaluate the selected indicators, prescribe how often the indicators should be reported to supplement SORTS data, and ensure that comparable data be maintained by the services to facilitate trend analysis.\3 DOD contracted the Logistics Management Institute (LMI) to study the indicators discussed in our report, and LMI found that 19 of them could be of high or medium value for monitoring critical aspects of readiness.\4 The LMI study, issued in 1994, recommended that DOD (1) identify and assess other potential indicators of readiness, (2) determine the availability of data to monitor indicators selected, and (3) estimate benchmarks to assess the indicators. Although our study and the LMI study concluded that a broader range of readiness indicators was needed, both left open how DOD could best integrate additional measures into its readiness reporting. The 19 indicators that Congress is requiring DOD to include in its quarterly reports are very similar to those assessed in the LMI study. (See app. 1 for a list of the 19 indicators DOD is to include in the quarterly reports.) Last month, DOD provided Congress with an implementation plan for meeting the expanded reporting requirements for the quarterly report. We were asked to comment on this plan today. Of course, a thorough assessment of the additional readiness indicators will have to wait until DOD begins to incorporate them into the quarterly reports in October 1998. However, on the basis of our review of the implementation plan, we have several observations to make. Overall, the implementation plan could be enhanced if it identified the specific information to be provided and the analysis to be included. The plan appears to take a step backward from previous efforts to identify useful readiness indicators. In particular, the LMI study and subsequent efforts by the Office of the Secretary of Defense were more ambitious attempts to identify potentially useful readiness indicators for understanding, forecasting, and preventing readiness shortfalls. The current implementation plan, in contrast, was developed under the explicit assumption that existing data sources would be used and that no new reporting requirements would be created for personnel in the field. Further, the plan states that DOD will not provide data for 7 of the 19 indicators because either the data is already provided to Congress through other documents or there is no reasonable or accepted measurement. DOD officials, however, acknowledged that their plans will continue to evolve and said they will continue to work with this Subcommittee to ensure the quarterly report supports congressional oversight needs. Lastly, the plan does not present a clear picture of how the additional indicators will be incorporated into the quarterly report. For example, the plan is mostly silent on the nature and extent of analysis to be included and on the format for displaying the additional indicators. We also have concerns about how DOD plans to report specific indicators. For example: -- According to the plan, SORTS will be the source of data for 4 of the 19 indicators--personnel status, equipment availability, unit training and proficiency, and prepositioned equipment. By relying on SORTS, DOD may miss opportunities to provide a more comprehensive picture of readiness. For example, the LMI study points out that SORTS captures data only on major weapon systems and other critical equipment. That study found value in monitoring the availability of equipment not reported through SORTS. In all, the LMI study identified more than 100 potential data sources outside SORTS for 3 of these 4 indicators--personnel status, equipment availability, and unit training and proficiency. (The LMI study did not include prepositioned equipment as a separate indicator.) -- DOD states in its implementation plan that 2 of the 19 indicators-- operations tempo (OPTEMPO) and training funding--are not relevant indicators of readiness. DOD states further it will not include the data in its quarterly readiness reports because this data is provided to Congress in budget documents,. However, the LMI study rated these two indicators as having a high value for monitoring readiness. The study stated, for instance, that "programmed OPTEMPO is a primary means of influencing multiple aspects of mid-term readiness" and that "a system for tracking the programming, budgeting, and execution of OPTEMPO would be a valuable management tool that may help to relate resources to readiness." -- For the indicator showing equipment that is non-mission capable, the plan states that the percentage of equipment reported as non-mission capable for maintenance and non-mission capable for supply will provide insights into how parts availability, maintenance shortfalls, or funding shortfalls may be affecting equipment readiness. According to the plan, this data will be evaluated by examining current non-mission capable levels versus the unit standards. While this type of analysis could indicate a potential readiness problem if non-mission capable rates are increasing, it will not show why these rates are increasing. Thus, insights into equipment readiness will be limited. -------------------- \3 Military Readiness: DOD Needs to Develop a More Comprehensive Measurement System (GAO/NSIAD-95-29, Oct. 27, 1994). \4 An Initial Assessment of GAO Collected Readiness Indicators: Their Value in Monitoring Mid-Term Readiness and Avoiding Hollowness, Logistics Management Institute (Oct. 1994). ARE FURTHER IMPROVEMENTS TO DOD'S REPORTING PROCESS NEEDED? ---------------------------------------------------------- Chapter 0:4 Mr. Chairman, there are two areas where we think DOD has an opportunity to take further actions to improve its readiness reporting. The first area concerns the level of detail included in the quarterly readiness reports to Congress. In a draft report we will issue later this month, we have recommended that the Secretary of Defense take steps to better fulfill the legislative reporting requirements under 10 U.S.C. 482 by providing (1) supporting data on key readiness deficiencies and (2) specific information on planned remedial actions in its quarterly readiness reports. As we discussed earlier, the quarterly reports we reviewed gave Congress only a vague picture of readiness. Adding more specific detail should enhance the effectiveness of the reports as a congressional oversight tool. DOD has concurred with our recommendation. The second area where DOD can improve its readiness reporting concerns DOD's plan to include additional readiness indicators in the quarterly report. The plan would benefit from the following changes: -- Include all 19 required indicators in the report. -- Make the report a stand-alone document by including data for all the indicators rather than referring to previously reported data. -- Further investigate sources of data outside SORTS, such as those reviewed in the LMI report, that could provide insight into the 19 readiness indicators. -- Develop a sample format showing how the 19 indicators will be displayed in the quarterly report. -- Provide further information on the nature and extent of analysis to be included with the indicators. DOD recognizes in its plan that the type and quality of information included in the quarterly reports may not meet congressional expectations and will likely evolve over time. In our view, it would make sense for DOD to correct known shortcomings to the current implementation plan and present an updated implementation plan to Congress prior to October 1998. -------------------------------------------------------- Chapter 0:4.1 Mr. Chairman, that concludes my prepared statement. We would be glad to respond to any questions you or other Members of the Subcommittee may have. ADDITIONAL READINESS INDICATORS TO BE INCLUDED IN THE QUARTERLY READINESS REPORT TO CONGRESS =========================================================== Appendix I The following are the additional indicators the Department of Defense is required, under 10 U.S.C. 482, to include in its quarterly reports to Congress beginning in October 1998. PERSONNEL STRENGTH --------------------------------------------------------- Appendix I:1 1. Personnel status, including the extent to which members of the armed forces are serving in positions outside of their military occupational specialty, serving in grades other than the grades for which they are qualified, or both. 2. Historical data and projected trends in personnel strength and status. PERSONNEL TURBULENCE --------------------------------------------------------- Appendix I:2 3. Recruit quality. 4. Borrowed manpower. 5. Personnel stability. OTHER PERSONNEL MATTERS --------------------------------------------------------- Appendix I:3 6. Personnel morale. 7. Recruiting status. TRAINING --------------------------------------------------------- Appendix I:4 8. Training unit readiness and proficiency. 9. Operations tempo. 10. Training funding. 11. Training commitments and deployments. LOGISTICS--EQUIPMENT FILL --------------------------------------------------------- Appendix I:5 12. Deployed equipment. 13. Equipment availability. 14. Equipment that is not mission capable. 15. Age of equipment. 16. Condition of nonpacing items. LOGISTICS--EQUIPMENT MAINTENANCE --------------------------------------------------------- Appendix I:6 17. Maintenance backlog. LOGISTICS--SUPPLY --------------------------------------------------------- Appendix I:7 18. Availability of ordnance and spares. 19. Status of prepositioned equipment. *** End of document. ***