1205-0431 Supporting Statement 071922

1205-0431 Supporting Statement 071922.docx

Unemployment Insurance Data Validation (DV) Program

OMB: 1205-0431

Document [docx]
Download: docx | pdf

Unemployment Insurance Data Validation (DV) Program

OMB Control No. 1205-0431

OMB Expiration Date: 9/30/2022


SUPPORTING STATEMENT FOR

Unemployment Insurance Data Validation (DV) Program

OMB CONTROL NO. 1205-0431


This Information Collection Request (ICR) seeks to extend without change the Unemployment Insurance Data Validation (DV) program.


  1. JUSTIFICATION


1. Explain the circumstances that make the collection of information necessary. Identify any legal or administrative requirements that necessitate the collection. Attach a copy of the appropriate section of each statute and regulation mandating or authorizing the collection of information.


The Unemployment Insurance Data Validation (DV) program assesses the accuracy of reports data which States submit to the Department monthly, quarterly or annually pursuant to the Department’s authority in Section 303(a)(6) of the Social Security Act.


The UI DV program validates 399 report elements from 16 UI required reports to ensure that these key data, used for performance management purposes, are reliable. Appendix A-1 contains the relevant sections of the Social Security Act on which the reporting and validation activity is based.


Section 303(a)(6) of the Social Security Act specifies that the Secretary of Labor will not certify State unemployment compensation programs to receive administrative grants unless the State’s law includes provision for: “...making of such reports...as the Secretary of Labor may from time to time require, and compliance with such provisions as the Secretary may from time to time find necessary to assure the correctness and verification of such reports.”


In the 1970s the Department discovered that the UI required reports data it was using to determine States’ administrative budget allocations were often reported inaccurately. In response, it developed the Workload Validation (WV) program to assess the accuracy of eleven key workload items used to allocate funds for UI administration among States. Under the WV program, States were required to validate 29 elements from four UI required reports. This validation activity was conducted using the instructions in the UI Workload Validation Handbook No. 361 approved for use through 12/31/2000 (OMB Number 1205-0055). Handbook No. 361 later became the UI Data Validation handbook, which is one of the instruments approved under 1205-0431.

The 29 elements validated through the WV program constituted slightly over 1% of the approximately 2,400 elements States report on 47 UI required reports and only a few Government Performance and Results Act (GPRA) elements. In light of this, the General Accounting Office and the DOL’s Office of Inspector General (OIG) recommended that the program be revised to cover more data elements associated with performance measures.


In the early 1990s, as part of a project to develop and test new measures for the timeliness and quality of benefit payment operations, called the Performance Measurement Review (PMR), the DOL asked the technical support contractor, Mathematica Policy Research (MPR), to develop a method for validating benefits data. MPR developed an approach based on WV concepts but more amenable to automation. It included nearly 1,200 data elements from 13 benefits reports. The methodology was subsequently extended to include the data reported on the primary tax performance report, ETA 581, “Contribution Operations” (OMB Control No. 1205-0178, expires 07/31/2024). After the method was tested in three states with satisfactory results, the DOL had MPR develop software that all states could use to perform validation operations. States installed this software on individual state PC or Sequel Server computers. After completing data validation, results could be exported from the software in different formats (e.g., text, Excel, Word, PDF files) and sent to the National Office via email.


Although the MPR software represented an improvement from the original DV methodology, DOL recognized that it could be enhanced, and as a response, DOL began developing its own version in 2003. The DOL software, released in 2005, is the software that is currently used to validate reported data. It has the advantage that it resides on the Sun-Unix servers that states use for UI reporting operations, which facilitates the deployment of new software versions, allows transmission of validation results directly to the National Office UI database, and allows automatic retrieval of reported counts from the UI database instead of requiring validators to enter those data as in the previous software. The latest version, Version 5.3.0, was released in February 2015.


The Department required states to implement UI DV. It set July 31, 2003 as the target for states to install the common UI DV software and develop the extracts of benefits and tax transactions the software processes; the states were to submit reports containing UI DV results by September 30, 2003. As of June 2015, reports had been received from 52 of the 53 states and territories that operate UI programs. All states and territories partially or fully complete data validation annually with the exception of the U.S. Virgin Islands.


Unemployment Insurance Program Letter (UIPL) 22-05, available on the DV Web site’s Advisories Archive at http://www.oui.doleta.gov/dv/archive_advisor.asp, summarizes the UI DV policy framework. Data Validation is administered on a “Validation Year” basis. The Validation Year (VY) coincides with the State Quality Service Plan (SQSP) performance year, which runs from April 1 through the following March 31. For example, a validation of any report for UI activity that occurs between April 1, 2021 and March 31, 2022 is part of VY 2022 and is to be submitted by June 10, 2022. The timing allows validation results to be part of the SQSP process.


The DV program operates on a three-year cycle. All but three groups of transactions—called populations--that pass validation must be validated every third year. The other three populations—which validate report cells from which GPRA indicators are formulated--must be validated annually; and populations that fail validation must be revalidated the next year. UI DV has also adopted the WV parameters for validity: if report counts are within ± 2% of reconstructed (validation) counts, and the random samples show that no more than 5% of transactions in the extract file universe contain invalid data elements, the reported counts are considered valid. (The GPRA report counts are held to a ± 1% standard instead of ± 2%.) The rules for passing validation are also contained in UIPL 22-05. Over 60% of all validation items due as of August 10, 2021, passed validation.

Both Benefits and Tax portions of the DV handbook are available on the ETA Office of Workforce Security web site at https://oui.doleta.gov/dv/index.asp


The DV handbook and software show the following approval information:


OMB No.: 1205-0431

OMB Expiration Date: 09/30/2022


OMB Burden Statements:  Response time for this collection of information is estimated to average 730 hours per response (this is the average of a full validation every third year with an estimated burden of 900 hours, and partial validations in the two intervening years), including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden to the U. S. Department of Labor, Employment and Training Administration, Office of Unemployment Insurance (Attn: Rachel Beistel), 200 Constitution Avenue, NW, Room S-4519, Washington, D.C. 20210 (Paperwork Reduction Project 1205-0431).


2. Indicate how, by whom, and for what purpose the information is to be used. Except for a new collection, indicate the actual use the agency has made of the information received from the current collection.


Accurately reported data are essential for properly assessing State performance and for ensuring that States are treated equitably when administrative funds are distributed. The DOL will use UI DV results to evaluate the accuracy of the data States report on selected UI required reports. This will enable it to assure its customers, partners and stakeholders of the accuracy of the performance measures that form the basis of the UI Performs system, allow the Congress to determine the extent the UI system is accomplishing the goals set for the GPRA Strategic and Annual Performance plans and also to ensure that UI administrative funds are allocated properly to the 53 State UI programs.


The DOL will use the UI DV information as a performance measure in the SQSP process. SQSP is a permanent framework for the annual planning and budget process in each State UI agency. As part of the SQSP process, states that inaccurately submit data on validated reports—from which, among other things, key UI Performs or GPRA measures are based--will be required to take corrective action to ensure accurate reporting.


Because the performance measures are intended to guide and assist States in improving their performance and service to their ultimate customers (workers and employers), States also have an interest in ensuring that the measures are correct. The UI DV system includes detailed guidance to State programmers for correctly constructing various report elements. This not only ensures they will construct the DV extract files correctly, but also will simplify their tasks when new tax or benefit system software is installed.


Prior to implementing UI DV, the Department was criticized for failing to validate the data on which key measures are based and for failing to increase the extent of the validation of the reports data it receives from States. The UI DV system allows the DOL to respond positively to those criticisms and to ensure the accuracy of the reports data on which it relies to discharge its oversight responsibilities.


3. Describe whether, and to what extent, the collection of information involves the use of automated, electronic, mechanical, or other technological collection techniques or other forms of information technology, e.g., permitting electronic submission of responses, and the basis for the decision for adopting this means of collection. Also, describe any consideration of using information technology to reduce burden.


In compliance with the Government Paperwork Elimination Act, beginning on April 1, 2005, the UI DV program has had States submit UI DV findings via the system now used for electronic transmission of UI reports.


The DOL has provided each State UI agency with an automated menu-driven system to input, store, and transmit reports data. The UI DV results are entered electronically as a report and stored in the UI Database.


The UI DV system is highly automated. The State-specific UI DV handbooks give States detailed specifications for developing data extract files, based on their own State management information systems. The States also enter or download these extract files to the automated system where a software module manipulates the input files for validation by producing validation displays, sorts, reconstruction counts, duplicate detection tests, and drawing various samples. The software module extracts from the state’s report database the report elements being validated so that validators do not need to enter those counts. The software also provides input interfaces where validators enter the findings of their reviews of quality samples, known as Data Element Validation. The results of completed validations are reported electronically to the Department using the overnight pick-up mechanism used to transmit other UI reports data.


In the last few years the DOL has increased UI DV automation. UI DV summary results from the DOL database are available to DOL and State users through a Web-based system. A separate Web-based system allows State users access to a database to maintain Module 3 of the benefits and tax handbooks. Module 3 is the “map” that enables the States to relate their key data to Federal reporting definitions. It gives Federal definitions from reporting and validation requirements; and based on those requirements, States note where these elements are to be found in their management information systems. State programmers use Module 3 to find the data elements needed to build DV extract files; validators use it to test the extract files when they examine sampled records during the data element validation phase.


The Department knows of no technical obstacles to implementing and operating UI DV.


4. Describe efforts to identify duplication. Show specifically why any similar information already available cannot be used or modified for use for the purposes described in Item A.2 above.


The proposed UI DV program does not duplicate any existing Department program. The Department believes that there is no satisfactory alternative to the UI DV methodology for this purpose.


5. If the collection of information impacts small businesses or other small entities, describe any methods used to minimize burden.


There is no impact on small businesses or any entities other than State UI agencies.


6. Describe the consequence to federal program or policy activities if the collection is not conducted or is conducted less frequently, as well as any technical or legal obstacles to reducing burden.


The Department requires that States conduct a complete validation every third year and a partial validation every year for those items determined to be reported inaccurately by a previous year’s validation results or that are used for GPRA indicators. A validation is required within a year after states change State reports-generating software, such as a new State management information system or the Department changes reporting requirements.


7. Explain any special circumstances that would cause an information collection to be conducted in a manner:


  • requiring respondents to report information to the agency more often than quarterly;


  • requiring respondents to prepare a written response to a collection of information in fewer than 30 days after receipt of it;


  • requiring respondents to submit more than an original and two copies of any document;


  • requiring respondents to retain records, other than health, medical, government contract, grant-in-aid, or tax records for more than three years;


  • in connection with a statistical survey, that is not designed to produce valid and reliable results that can be generalized to the universe of study;


  • requiring the use of statistical data classification that has not been reviewed and approved by OMB;


  • that includes a pledge of confidentiality that is not supported by authority established in statute or regulation, that is not supported by disclosure and data security policies that are consistent with the pledge, or which unnecessarily impedes sharing of data with other agencies for compatible confidential use; or


  • requiring respondents to submit proprietary trade secret, or other confidential information unless the agency can demonstrate that it has instituted procedures to protect the information's confidentiality to the extent permitted by law.


No such circumstances exist. This request is consistent with 5 CFR 1320.5.


8. If applicable, provide a copy and identify the date and page number of publication in the Federal Register of the agency's notice, required by 5 CFR 1320.8(d), soliciting comments on the information collection prior to submission to OMB. Summarize public comments received in response to that notice and describe actions taken by the agency in response to these comments. Specifically address comments received on cost and hour burden.


Describe efforts to consult with persons outside the agency to obtain their views on the availability of data, frequency of collection, the clarity of instructions and recordkeeping, disclosure, or reporting format (if any), and on the data elements to be recorded, disclosed, or reported.


Consultation with representatives of those from whom information is to be obtained or those who must compile records should occur at least once every 3 years -- even if the collection-of-information activity is the same as in prior periods. There may be circumstances that may preclude consultation in a specific situation. These circumstances should be explained.


In accordance with the Paperwork Reduction Act of 1995, the public was allowed 60 days to comment through the Federal Register notice posted on January 24, 2022 (87 FR 3588). No public comments were received.


9. Explain any decision to provide any payments or gifts to respondents, other than remuneration of contractors or grantees.


This program does not involve payments per se to respondents. However, because UI DV is a mandated data collection, the DOL provides funding to the participating States who are technically listed as the "respondents" for purposes of the Paperwork Reduction Act. States undertake validation activities out of their UI administrative grant funds, the same process for its predecessor, workload validation.


10. Describe any assurance of confidentiality provided to respondents and the basis for the assurance in statute, regulation, or agency policy.


Keeping data private is not an issue with this program, which simply involves verifying the accuracy of aggregated data counts, submitted earlier on required reports. States submit no individual records with personal (e.g. Social Security Numbers) or business (e.g., Employer Account Number) identifiers.


11. Provide additional justification for any questions of a sensitive nature, such as sexual behavior and attitudes, religious beliefs, and other matters that are commonly considered private. This justification should include the reasons why the agency considers the questions necessary, the specific uses to be made of the information, the explanation to be given to persons from whom the information is requested, and any steps to be taken to obtain their consent.


Not applicable.


12. Provide estimates of the hour burden of the collection of information. The statement should:


  • Indicate the number of respondents, frequency of response, annual hour burden, and an explanation of how the burden was estimated. Unless directed to do so, agencies should not conduct special surveys to obtain information on which to base hour burden estimates. Consultation with a sample (fewer than 10) of potential respondents is desirable. If the hour burden on respondents is expected to vary widely because of differences in activity, size, or complexity, show the range of estimated hour burden, and explain the reasons for the variance. Generally, estimates should not include burden hours for customary and usual business practices.


  • If this request for approval covers more than one form, provide separate hour burden estimates for each form.


  • Provide estimates of annualized cost to respondents for the hour burdens for collections of information, identifying and using appropriate wage rate categories. The cost of contracting out or paying outside parties for information collection activities should not be included here. Instead, this cost should be included in Item 14.




Annual Hour Burden: 23,638 hours

Old Burden: 23,644 hours

Change: -6

  • Number of respondents: 53 State Employment Security Agencies

  • Burden for a full validation: 733 hours

    • Based on reassessed estimates received from three States that have conducted validations; they assumed they were conducting routine validations for a normal validation year using the validation software.

  • Annual Estimate Assumes:

    • States conduct a full validation in year 1;

    • In year 2, they validate half of report cells (i.e., GPRA elements and some failures from full validation, based on initial validation results that show at least half of report items failed);

    • In year 3, they validate one third of report cells in third year (i.e., the GPRA populations and some continued failures from validation in year 2);

    • These assumptions may be conservative; degree of improvement from one validation to another is not known. Burden is expected to decline in the future as report validity improves due to fewer report cells to be revalidated and more random samples passing at the first stage.

  • Calculation of Average Annual Burden

    • Full validation = 53*730 = 38,690

    • hours (Year 1)

    • Half validation = 53*730/2 = 19,345 (Year 2)

    • Third validation = 53*730/3 = 12,897 (Year 3)

    • 3-Year Average = (38,690 + 19,345 + 12,879)/3 = 23,638


  • Annual Burden = 23,638 hours

  • Average Wage Rate for State UI agencies: $53.27 = average salary used for UI budget formulation purposes for FY 2022 = $91,144 average yearly work hours used is 1,710)

  • Annual Monetized Value of the Hour Burden = 23,638 hours * $53.27/hour = $1,259,196

  • Change: $-320


Startup Burden to Respondents: N/A

All states have implemented DV.


The following table can be used as a guide to calculate the total burden of an information collection. Based on anticipated passing rates and that most DV scores are valid for three years, the average burden and annual cost estimates below are based on these assumptions:

  • Year 1 states: most benefits and tax populations are due

  • Year 2 states: about half of all benefits and tax populations are due

  • Year 3 states: about one-third of all benefits and tax populations are due



Activity

Number of Respondents

Number of Responses per Respondent

Total Responses

Average

Annual Burden (Hours)

Total Annual Burden (Hours)

Hourly Rate*

Monetized Value of Respondent Time

Unemployment Insurance Data Validation Program

53

1

53

446

23,638

$53.27

$1,259,196

  • *Source: The hourly rate is computed by dividing the FY 2022 national average PS/PB annual salary for state staff as provided for through the distribution of state UI administrative grants (https://wdr.doleta.gov/directives/corr_doc.cfm?DOCN=6102)by the average number of hours worked in a year (1,711).  For FY 2022, this calculation is:  $91,144 / 1,711= $53.27.


13. Provide an estimate of the total annual cost burden to respondents or record keepers resulting from the collection of information. (Do not include the cost of any hour burden shown in Items 12 and 14).


  • The cost estimate should be split into two components: (a) a total capital

and start up cost component (annualized over its expected useful life); and (b) a

total operation and maintenance and purchase of service component.

The estimates should take into account costs associated with generating,

maintaining, and disclosing or providing the information. Include descriptions of

methods used to estimate major cost factors including system and technology acquisition, expected useful life of capital equipment, the discount rate(s), and the time period over which costs will be incurred. Capital and start-up costs include, among other items, preparations for collecting information such as purchasing computers and software; monitoring, sampling, drilling and testing equipment; and record storage facilities.


  • If cost estimates are expected to vary widely, agencies should present ranges of cost burdens and explain the reasons for the variance. The cost of purchasing or contracting out information collection services should be a part of this cost burden estimate. In developing cost burden estimates, agencies may consult with a sample of respondents (fewer than 10), utilize the 60-day pre-OMB submission public comment process and use existing economic or regulatory impact analysis associated with the rulemaking containing the information collection, as appropriate.


  • Generally, estimates should not include purchases of equipment or services, or portions thereof, made: (1) prior to October 1, 1995, (2) to achieve regulatory compliance with requirements not associated with the information collection, (3) for reasons other than to provide information or keep records for the government, or (4) as part of customary and usual business or private practices.


There are no costs to respondents. The Department provides States with the computer equipment necessary for retrieving, manipulating, storing, and reporting the validation results.


14. Provide estimates of the annualized cost to the Federal Government. Also, provide a description of the method used to estimate cost, which should include quantification of hours, operational expenses (such as equipment, overhead, printing, and support staff), any other expense that would not have been incurred without this collection of information. Agencies also may aggregate cost estimates from Items 12, 13, and 14 into a single table.


There are no costs to the Federal Government.


15. Explain the reasons for any program changes or adjustments.


There are no program changes just adjustments due to estimates as reported in section 12.


16. For collections of information whose results will be published, outline plans for tabulations, and publication. Address any complex analytical techniques that will be used. Provide the time schedule for the entire project, including beginning and ending dates of the collection of information, completion of report, publication dates, and other actions.


The DOL publishes results of validations on its DV Website at https://oui.doleta.gov/dv/index.asp.


17. If seeking approval to not display the expiration date for OMB approval of the information collection, explain the reasons that display would be inappropriate.


The Department displays approval information in the ETA Handbook 361.


18. Explain each exception to the certification statement.


There are no exceptions.




B. COLLECTIONS OF INFORMATON EMPLOYING STATISTICAL METHODS.

See attached.

1


File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
File TitleMarch 6, 2007
AuthorTheda Kenney
File Modified0000-00-00
File Created2022-07-27

© 2024 OMB.report | Privacy Policy