Language selection

Search


Iterative design for AI help applications: Canada.ca design

Learning and feedback are crucial for successful agile service development. With the rapid pace of generative AI development, you must experiment and learn quickly. Testing should be an ongoing activity that spans the entire service design lifecycle. It should include not only functionality checks but also a wide range of factors including user experience, features of the tool, accessibility, official language adherence, inclusivity, unintended biases, etc.

On this page

A key aspect of this iterative design approach is identifying and addressing user pain points in accessing services. Programs should focus particularly on areas where the administrative burden is high, such as when users face significant learning, psychological, or compliance costs. This means understanding the complexities surrounding eligibility for services, the steps required to access them, and the general user experience.

We propose experimentation phases that include user research and testing throughout the entire lifecycle. Usability research helps to ensure that by the time of a full public launch, the application design meets user expectations and desired outcomes. Continuous evaluation of accuracy, security and usability should also be integrated into the design process, including testing with assessors, accessibility experts, and actual users to ensure it continues to meet expectations.

Actively identifying user pain points can also help inform iterative improvements and streamline access to services, ultimately creating a more client-centric experience. The GC Digital Standards contains information about designing with users and what it means to iterate and improve frequently.

Monitor and iterate

User research in the design phase

Involve user experience (UX) researchers early so they can start testing design prototypes with potential end users. Apply best practices for federal institutions deploying a generative AI tool from the Treasury Board of Canada Secretariat’s:

Use evaluation datasets from Canada.ca feedback

Every weekday, users leave up to 4,000 feedback comments across the site, many in the form of questions. Consult your web team or the Canada.ca Experience Office to generate an evaluation dataset from this user feedback. Before using the dataset, ensure you remove all comments with redacted personal information (comments that include the # character).

For more information on using feedback, see:

Initial testing phase

Provide call centre team members or other stakeholders with early versions of the application. They can provide feedback and rate/evaluate results.

Iterative usability testing sessions

As the design process continues, UX researchers should continue to test various aspects and iterations of the design with end-users to ensure it aligns with their needs and expectations. Unmoderated video software works well for this.

Perform as many iterations as needed to refine the product to meet client-centric standards.

The Guideline on Service and Digital provides additional information on client-centric service design and delivery:

Measure outcomes using feedback mechanisms and a post-use survey

Generative AI is rapidly changing expectations, and underlying models can evolve too. It’s crucial that AI chat applications deliver answers that meet users’ needs and expectations. You must continually manage and improve the accuracy and trustworthiness of these applications on Canada.ca. Feedback mechanisms can provide both quantitative and qualitative data for iterative improvement and promptly highlight issues that require immediate attention.

Drawing from research by GOV.UK, we recommend including a line below the answer that states, “How was this answer? Good or Needs improvement.” This phrasing better aligns with client-centric standards and invites users to provide meaningful feedback.

If a user selects “Needs improvement” consider offering follow-up questions to clarify their experience, such as:

The Guideline on Service and Digital provides additional information on client feedback:

Task Success Survey

On Canada.ca, we randomly invite visitors to respond to the Task Success Survey (TSS). We use the results to evaluate and improve success by category.

We recommend adding a special version of the TSS to your AI application. To compare the success of visitors using your AI application with those using only the web pages, you could display the TSS invitation to a random selection of AI users. Additionally, you could always have the TSS link available at the bottom of the application page.

The Canada.ca Experience Office has created and usability tested a version of the TSS for AI applications. You can access the survey here: https://cdssnc.qualtrics.com/jfe/form/SV_4N2YTcAHkcBEGfs

Teams should use survey data and evaluations of chat interactions to inform their decisions and determine how effective the help application is at delivering the desired service outcomes. Success is measured not by the number of people using the application, but by how many report that the application solved their problem and the rated accuracy of the answers they received.

For more information, see the guidance on using the GC Task Success Survey.

Page details

Date modified: