SUP STAT (1545-2290) (updated)

SUP STAT (1545-2290) (updated).docx

Improving Customer Experience (OMB Circular A-11, Section 280 Implementation)

OMB: 1545-2290

Document [docx]
Download: docx | pdf



SUPPORTING STATEMENT

Internal Revenue Service

Improving Customer Experience (OMB Circular A-11, Section 280 Implementation)

OMB Control No. 1545-NEW




1. Abstract


On September 11, 1993, President Clinton issued Executive Order 12862, “Setting Customer Service Standards” which clearly define his vision that the Federal agencies will put the people first. Executive Order 12862 directs Federal agencies to provide service to the public that matches or exceeds the best service available in the private sector. Section 1(b) of Executive Order 12862 requires government agencies to “survey customers to determine the kind and quality of services they want and their level of satisfaction with existing services” and Section 1(e) requires agencies “survey front-line employees on barriers to, and ideas for, matching the best in business.”


On March 30, 2016, President Obama established the Core Federal Services Council, which again emphasized the need to deliver world-class customer service to the American people. The Council, composed of the major high-volume, high-impact Federal programs that provide transactional services directly to the public, were encouraged “to improve the customer experience by using public and private sector management best practices, such as conducting self-assessments and journey mapping, collecting transactional feedback data, and sharing such data with frontline and other staff.”


In March 2018, the Administration of President Trump launched the President’s Management Agenda (PMA) and established new Cross-Agency Priority (CAP) Goals. Excellent service was established as a core component of the mission, service, stewardship model that frames the entire PMA, embedding a customer-focused approach in all of the PMA’s initiatives. This model was also included in the 2018 update of the Federal Performance Framework in Circular A-11, ensuring ‘excellent service’ as a focus in future agency strategic planning efforts. The PMA included a CAP Goal on Improving Customer Experience with Federal Services, with a primary strategy to drive improvements within 25 of the nation’s highest impact programs. This effort is supported by an interagency team and guidance in Circular A-11 requiring the collection of customer feedback data and increasing the use of industry best practices to conduct customer research.


These Presidential actions and requirements establish an ongoing process of collecting customer insights and using them to improve services. This new request will enable the Internal Revenue Service (hereafter “the Agency”) to act in accordance with OMB Circular A-11 Section 280 to ultimately transform the experience of its customers to improve both efficiency and mission delivery and increase accountability by communicating about these efforts with the public.

2. Use of Data


The Agency will collect, analyze, and interpret information gathered through this generic clearance to identify services’ accessibility, navigation, and use by customers, and make improvements in service delivery based on customer insights gathered through developing an understanding of the user experience interacting with Government.


For the purposes of this request, "customers" are individuals, businesses, and organizations that interact with a Federal Government agency or program, either directly or via a Federal contractor.


"Service delivery" or "services" refers to the multitude of diverse interactions between a customer and Federal agency such as applying for a benefit or loan, receiving a service such as healthcare or small business counseling, requesting a document such as a passport or social security card, complying with a rule or regulation such as filing taxes or declaring goods, utilizing resources such as a park or historical site, or seeking information such as public health or consumer protection notices.


Under this request, three types of activities will be conducted to generate customer insights:


Customer Research (User Persona and Journey Map Development): A critical first component of understanding customer experience is to develop customer personas and journey maps. This process enables the Agency to more deeply understand the customer segments they serve and to organize the processes customers interact with throughout their engagement with the Federal entity to accomplish a task or meet a need. In order to adequately capture the perspective of the customer and the barriers or supports that exist as they navigate these journeys, it is necessary to directly interact with customers rather than relying solely upon the Agency’s stated policy of how a process should work or employees’ interpretation of how services are delivered. This can occur through a variety of information collection mechanisms that include focus groups, individual intercept interviews at a service site, shadowing a user as they navigate a Federal service and documenting their reactions and frustrations, customer free-response comment cards, or informal small discussion groups.


Regardless of the format, the Agency will apply Human Centered Design (HCD) Discovery methods to generate personas and journey maps, ultimately identifying customer insights. An approach to recruiting participants, resources for preparing and structuring interviews, and a consent form for interviewees can be found at https://www.gsa.gov/cdnstatic/HCD-Discovery-Guide-Interagency-v12-1.pdf. This document is also included in the package.




Insights documented, summarized and presented in customer personas and journey maps can then be shared across the program, the Agency, other Federal, State, and Local government stakeholders and even with the public to validate and discuss common themes identified. These products can be used as “indicator lights” for where more rigorous qualitative and quantitative research can be conducted to improve Federal service delivery.


Publicly shared personas and journey maps will include language that qualifies their use (see question #16), and high-level, non-identifying descriptive statistics of the population(s) interviewed to develop it (ex. “25 Service members that transitioned to civilian employment within the last decade, 14 female, 11 male, 21 enlisted and 4 officers) to ensure that the perspective represented is understood. Quotes or insights will never be associated with an actual individual unless they have signed a release form (see link above for template) and this was included in the specific collection request.


Customer Feedback (Satisfaction Survey): Surveys to be considered under this generic clearance will only include those surveys modeled on the OMB Circular A-11 CX Feedback survey to improve customer service by collecting feedback at a specific point during a customer journey. This could include upon submitting a form online on a Federal website, speaking with a call center representative, paying off a loan, or visiting a Federal service center.


In an effort to develop comparable, government-wide scores that will enable cross-agency or industry benchmarking (when relevant) and a general indication of an agency’s overall customer satisfaction, OMB Circular A-11 Section 280 requires high impact services to measure their touchpoint/transactional performance in as a real-time manner as possible, with respect to satisfaction and confidence/trust using the following questions, without modification. Responses will be assessed on a 5-point Likert scale (1 (strongly disagree) to 5 (strongly agree)). These questions were developed in consultation with leading organizations in customer experience both in the private sector and industry groups that study the most critical drivers of customer experience.

  • 5 point Likert scale: I am satisfied with the service I received from [Program/Service name].

  • 5 point Likert scale: This interaction increased my confidence in [Program/Service name]. OR I trust [Agency/Program/Service name] to fulfill our country’s commitment to [relevant population].

  • Free response: Any additional feedback on your scores above?

  • 5 point Likert scale: My need was addressed OR My issue was resolved. OR I found what I was looking for.

  • 5 point Likert scale: It was easy to complete what I needed to do.

  • 5 point Likert scale: It took a reasonable amount of time to do what I needed to do.

  • 5 point Likert scale: I was treated fairly.

  • 5 point Likert scale: Employees I interacted with were helpful.

  • Free response: Any additional feedback for [Program/Service name]?




The surveys shall include no more than 15 questions in total. The Agency may add a few additional questions to those listed above to clarify type of service received, inquiry type, service center location, or other program-specific questions that can help program managers to filter and make use of the feedback data.


As part of the Customer Experience CAP goal’s strategy to increase transparency to drive accountability, the feedback data collected through the A-11 Standard Feedback survey is meant to be shared with the public. This collection is part of the government-wide effort to embed standardized customer metrics within high-impact programs to create government-wide performance dashboards. Data collected from the questions listed above will be submitted by the Agency to OMB quarterly for updating of customer experience dashboards on performance.gov. This dashboard will also include the total volume of customers that passed through the transaction point at which the survey was offered, the number of customers the survey was presented to, the number of responses, and the mode of presentation and response (online survey, in-person, post-call touchtone, mobile, email). This will help to qualify the data’s representation by showing both the response rate and total number of actual responses.


User Testing of Services and Digital Products: Agencies should continually review, update and refine their service delivery, including communication materials, processes, supporting reference materials, and digital products associated with a Federal program. This often requires “field testing” program informational materials, process updates, forms, or digital products (such as websites or mobile applications) by interacting with past, existing, or future customers and soliciting feedback. These activities can include cognitive laboratory studies, such as those used to refine questions on a program form to ensure clarity, demo kiosks at a service center where customers can provide informal feedback while waiting for a service, or more formally scheduled in-person observation testing (e.g., website or software usability tests). These information collection activities are more specific than broad customer research and related to a particular artifact / product of a Federal program. As such, there will be a more structured interview / set of questions than more open-ended customer research. Findings from these activities are meant to support the design and implementation of Federal program services and digital products, and may only be shared in an anonymized / in aggregate if a particular insight is useful to include as part of a customer persona, journey map, or common lesson learned for improving service delivery.


The Agency will only submit under this generic clearance if it meets the following conditions:


  • The collections are voluntary;

  • The collections are low-burden for respondents (based on considerations of total burden hours or burden-hours per respondent) and are low-cost for both the respondents and the Federal Government;

  • The collections are non-controversial and do not raise issues of concern to other Federal agencies;

  • Any collection is targeted to the solicitation of opinions from respondents who have experience with the program or may have experience with the program in the near future;

  • Personal identifiable information (PII) is collected only to the extent necessary and is not retained;

  • Information gathered is intended to be used for general service improvement and program management purposes

  • Upon agreement between OMB and the agency all or a subset of information may be released as part of A-11, Section 280 requirements only on performance.gov. Additionally, summaries of customer research and user testing activities may be included in public-facing customer journey maps.

  • Additional release of data must be done coordinated with OMB.


This clearance will help the Agency to establish a process where customer experience is regularly monitored and measured. The results will assist the Agency in the planning and decision-making processes to improve the quality of the Agency’s products and services.


Results from feedback activities and surveys will be used to measure against established baseline standards and for measuring the Agency’s progress toward defined goals.


3. Use of Improved Information Technology to Reduce Burden


There are neither legal nor technical obstacles to the use of technology in these information collection activities. The determination to use technology, and which technology to use, will be based on the type of information collected and the utility and the availability of specific technology to each respondent in a proposed customer research activity or feedback survey.


4. Efforts to Identify Duplication


The Agency will work to ensure the streamlining of all customer research and feedback surveys under this clearance. The Agency will also work to reduce existing customer feedback surveys and questions into alignment with the A-11 Standard CX Feedback survey as part of a coordinated Agency-wide customer program. The information to be supplied on these surveys will not be duplicated on any other information collection.


5. Methods to Minimize Burden on Small Businesses or Other Small Entities


The information collected in these surveys will represent the minimum burden necessary to evaluate customer experience with the Agency’s programs and processes. The Agency will minimize the burden on respondents by sampling as appropriate, asking for readily available information, and using short, easy-to-complete information collection instruments.



6. Consequences of Less Frequent Collection on Federal Programs or Policy Activities


Without regular mechanisms for collecting and generating customer insights, the Agency is not able to provide the public with the highest level of service. These activities will be coordinated to ensure that most individual respondents will not be asked to respond to more than one survey instrument per transaction or to participate in more than one qualitative feedback or testing activity.


7. Special Circumstances Requiring Data Collection to be Inconsistent with Guidelines in 5 CFR 1320.5(d)(2)


These surveys will be consistent with all the guidelines in 5 CFR 1320.5, especially those provisions in subsection (g) which require that a statistical survey be designed to produce results that can be generalized to the universe of study. There are no special circumstances that would cause this information collection to be conducted in an unusual or intrusive manner. All participation will be voluntary. Should the Agency need to deviate from the requirements outlined in 5 CFR 1320, individual justification will be provided to OMB on a case-by-case basis.


  1. Consultation with Individuals Outside of the Agency on Availability of Data, Frequency of Collection, Clarity of Instructions and Forms, and Data Elements


In response to the Federal Register notice dated April 11, 2023 (88 FR 21748), IRS

received no comments during the comment period regarding this collection.


9. Explanation of Decision to Provide Any Payment or Gift To Respondents


The standard will be no payment or gift to respondents for participation. If any payments are proposed the Agency will submit specific justification for each proposed use as part of the completed package submitted to OMB.


10. Assurance of Anonymity of Responses


Activity and survey instructions will provide all necessary assurances of confidentiality to the respondents. Although there is no requirement for such an assurance in statute, the quality of this type of information requires respondent candor and anonymity.


11. Justification of Sensitive Questions



No questions will be asked that are of a personal or sensitive nature.






12. Estimated Burden of Information Collection


A variety of instruments and platforms will be used to collect information from respondents. The annual burden hours requested (150,000) is based on the number of collections we expect to conduct over the requested 3-year approval period for this clearance.

Customer Interviews: The respondent burden averages from five (5) minutes to one hundred and twenty (120) minutes for interviews or focus groups, dependent on the individual activity. This estimation has been based on previous customer research activities conducted by Federal customer experience teams. The Agency estimates that across all sub-components and services, there will be up to 2,232 individual interviews averaging 2 hours each.

Focus Groups: The Agency estimates that there will be approximately 10,713 potential participants. The Agency estimates that there will be 50 segmented focus groups held each year with an average of thirty participants per group for. This clearance estimates that each member of a focus group is expected to spend an average of 1.5 hours per group.

Customer Feedback (Satisfaction Survey): Industry best practice is to present every customer the opportunity to provide feedback (8 minutes or less) at each instrumented touchpoint/transaction in a customer journey (ex. After submitting an application, completing a call at a call center, or visiting an in-person service center). The Agency will specify the total possible number of respondents based on estimated annual volume, but this information collection sets a ceiling estimate of 994,336 annually.


Testing of Services and Digital Products: The Agency estimates that there may be approximately 2,232 potential respondents. Simple user testing (30 minutes or less) can be useful for a variety of digital product and service initiatives across the Agency (ex. basic website design updates, communications revisions). Expecting each instance to have on average of 20 users, and 75 services/products tested each year.


User Testing: Additionally, more extensive cognitive laboratory or in-person observations can occur for newly launched products (such as mobile applications or re-designs of process steps) requiring an average of 30 minutes per respondent. Expecting each instance to have an average of 30 users, and 33 new services/products tested each year across the Agency. The Agency estimates that this adds 1,488 respondents to the clearance.









The Agency will keep track of the above activities in order to accurately update burden calculations year to year.

Type of Information Collection

Number of Respondents

Responses (per respondent)

Time per response (approx.)

Total Burden hours

Customer Interviews

2,232

1

2 hours

4,320

Focus Groups

10,713

1

90 minutes

15,554

Feedback Survey

994,336

1

8 minutes

128,326

User Testing (Rapid Feedback)

2,232

1

30 minutes

1,080

User Testing (Deep Dive)

1,488

1

30 minutes

720

Total

1,011,000



150,000






13. Estimated Total Annual Cost Burden to Respondents


No costs for respondents are anticipated.


14. Estimated Annualized Cost to the Federal Government


The Agency may incur costs in setting up focus group, small discussion, or testing environments to include such things as hiring contractors, facilitators or moderators, travel to customer sample locations, renting meeting space, in providing remuneration, etc. Costs will be determined on an individual project basis and will be included in the ICR provided to OMB for each project to be conducted.


15. Reasons for Change in Burden


There are no changes to the scope of the generic clearance. IRS is making an update to request additional burden hours. This is increasing the burden hours by 45,845 hours due to Agency Estimate.


16. Plans for Tabulation, Statistical Analysis and Publication


No attempt will be made to generalize the findings from these three groups of activities to be nationally representative or statistically valid. They are meant to compliment and help to contextualize performance and evaluation data as part of a three-pronged approach to understanding Federal program implementation and opportunities for improvement (Performance, Evaluation, and “Feedback” data1).


Customer Research: Insights gleaned from qualitative customer research may be presented publicly in the format of a conceptual user persona or customer journey map. Customer research can take anywhere from 6 weeks for a short sprint to a full fiscal year, depending on the specific project. The Agency expects most journey mapping efforts to last approximately 6 months, with a user persona and journey maps ready for feedback (both from internal and external to government stakeholders) within one month of completing customer research.

Publicly available Journey maps will include specific language to contextualize their use and will be included in specific requests. This language can include something like:


What should I know about journey maps?

Journey maps are living documents—continually refined and revisited. There is never

a “final” version, and these maps are meant to serve as a summary of the voices of

actual customers of U.S. Government services. A map may not precisely document

the way a Government program is meant to be navigated, accessed, or used. It might

not capture every government program or resource available to a customer segment.

However, it is the product of a qualitative research approach to gather insights from

customers’ actual experiences. These findings can help us to identify areas for

high-impact improvements across delivery channels and organizational silos.


Customer Feedback: Once touchpoint surveys are implemented at transaction points along the customer journey interacting with Federal services, data from the A-11 Standard CX Feedback survey will be submitted to OMB quarterly for review and publication in a summary dashboard on performance.gov.


This data will include:

  • Specific transaction point at which the survey was administered

  • Total volume of customers that interacted at this transaction point during the given quarter

  • Total volume of customers that were presented the survey

  • Total number of customers who completed the survey

  • Mode(s) of collection (ex. online, over mobile, over the phone, paper form)

  • Specific survey instrument that shows the Agency’s wording of standard A-11 CX Feedback survey

  • Distribution of the responses across the 5-point Likert scale for each of the standard questions


The purpose of collecting volume and response numbers is to share customer feedback measures in context of the response rate and total volume of responses to qualify interpretation of the CX feedback data.


Testing of Services and Digital Products: Similar to Customer Research, this can range from a short two-day rapid feedback from users within an Agile product development sprint or longer effort to gather more extensive feedback from multiple physical locations.


17. Reasons Why Displaying the OMB Expiration Date Is Inappropriate


Displaying the expiration date may cause problems with respondents for data collection programs that overlap the three-year approval periods. Would-be respondents might be inclined to refuse to participate if the form carries a date that is expired or is soon to expire.


18. Exceptions to the Certification Statement


There are no exceptions to the certification statement for this collection.

Note: The following paragraph applies to all the collections of information in this submission:

An agency may not conduct, or sponsor and a person is not required to respond to a collection of information unless the collection of information displays a valid OMB control number. Books or records relating to a collection of information must be retained if its contents may become material in the administration of any internal revenue law. Generally, tax returns and tax return information are confidential, as required by 26 U.S.C. 6103.





Shape1

10


File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
File TitleSUPPORTING STATEMENT
AuthorKenneth Smith
File Modified0000-00-00
File Created2023-07-29

© 2024 OMB.report | Privacy Policy