For the best reading experience, please switch to desktop view 💻✨

Harness AI to Build AI:
An LLM-powered co-pilot that provides in-session support for tutors

Harness AI to Build AI:
An LLM-powered co-pilot that provides in-session support for tutors

Project Info

001

Timeline

Timeline

Ongoing

Role

UX Designer

Team

Zhiyuan Chen
Wesley Deng ()
Jason Hong ()

Methods

Focus Group

Company Info

002

Who is WeAudit?

Overview

003

Background:
Bias Beneath the AI Hype

AI is part of our everyday lives, from virtual assistants to customer service. While it’s praised for solving problems and improving efficiency, its hidden flaws, like bias, are often overlooked. Bias in AI can damage trust, lead to bad decisions, and create legal risks for businesses. On a broader level, it can reinforce stereotypes or exclude marginalized groups, worsening existing inequalities.

78%

of executives believe data bias will become a bigger problem as AI/ML use increases

13%

of AI specialists rated their AI projects as very fair

36%

of organizations surveyed reported suffering from AI bias, with 62% experiencing lost revenue and 61% losing customers as a result

Problem:
AI teams struggle to scaffold end user bias audits

AI practitioners in the industry have their cultural blind spots when auditing biases in AI systems. While they are highly motivated to involve end users in the auditing process, they lack a platform for end users' auditing process and struggle with providing the necessary context, support, and scaffolding to empower end users to identify and analyze biases meaningfully and effectively.

Solution:
Empowering Industry AI Product Team with an auditing platform

We created WeAudit, a crowdsourced platform that empowers users to identify, discuss, and report bias in AI-generated content. By offering a centralized space, WeAudit educates users on bias, fosters meaningful discussions, and encourages actionable insights to mitigate the negative impacts of AI bias.

Impact:
Endorsements from Tech Giants

WeAudit demo is presented to many AI teams from Google, Apple, Meta, Microsoft and have received positive feedback. Many expressed an interest in adopting similar features to improve their own AI systems, validating WeAudit’s approach to addressing bias.

Challenge before I join

004

Siloed Team, Disconnected solutions

Our team has many ideas in supporting users to audit bias in AI, and each idea is prototyped into an individual product. Each team runs in silos and they are very heads down on developing the product and expanding capabilities and creating new products.

Product 1️⃣

Users enter two prompts, generate 2 sets of images, and compare them to identify biases, e.g., “People wearing hats” vs. “Koreans wearing hats.”

Users enter two prompts, generate 2 sets of images, and compare them to identify biases, e.g., “People wearing hats” vs. “Koreans wearing hats.”

Users enter two prompts, generate 2 sets of images, and compare them to identify biases, e.g., “People wearing hats” vs. “Koreans wearing hats.”

Product 2️⃣

Users enter a prompt to generate images, the system runs a data visualization on the images to see distribution of race, gender, age, etc.

Users enter a prompt to generate images, the system runs a data visualization on the images to see distribution of race, gender, age, etc.

Users enter a prompt to generate images, the system runs a data visualization on the images to see distribution of race, gender, age, etc.

Product 3️⃣

Users enter a prompt and compare images generated by different text-to-image models to identify bias and report them

Users enter a prompt and compare images generated by different text-to-image models to identify bias and report them

Users enter a prompt and compare images generated by different text-to-image models to identify bias and report them

There are many more products…

Frustrating user experience and overwhelmed AI teams

Having different products pose many challenges for AI practitioners. First of all, the disconnected experience can be frustrating and discourage end users from using them. Second, all the reports and data are also separate, which creates additional effort for AI practitioners to aggregate and analyze to obtain meaningful insights.

Define MVP Scope

007

Stepping up to create a clear and focused MVP

I took the initiative to refocus the team by stating, “We need to stop adding more features and define what truly matters for an MVP.” To make this happen, I first consolidated all standalone products into a single platform. Then, I prioritized key features and deprioritized others, ensuring the MVP was both streamlined and impactful.

Finding the overlap

Each product came with its own set of features, which made integration complex. To simplify, I created a union set of all features across the products and identified the overlapping ones, which became the foundation of our MVP.

Choosing what matters the most

There were many features outside the union set that I evaluated using an impact-value matrix. I assessed each feature by balancing its value, technical feasibility, and usability for users to determine whether it should be included in the MVP.