Voice typing

Make feedback more inclusive by empowering users to give feedback entirely through the modality of their preference.

Make feedback more inclusive by empowering users to give feedback entirely through the modality of their preference.

Focus areas

Interaction design

User research

Collaboration

Context

Google Feedback is a core primitive used across products to let users report issues and share ideas. However, after analyzing large volumes of feedback, we found that users were also using it to seek help, vent frustrations, or just type gibberish—causing valuable insights to get lost in the noise. We took a step back to rethink the end-to-end feedback journey with the following goals:

  1. Empower teams to configure feedback intake based on their product context.

  2. Reduce junk volume by helping users articulate better and redirecting them to appropriate channels.

  3. Enable teams with AI-powered theme detection and channels for user updates when issues are resolved.

  4. Acknowledge user effort to encourage better feedback over time.

Orientation - the feedback journey

Challenges & opportunities

Solutions

➡️

Product teams set up feedback as per their requirements

😔

Limited customizations

😔

Heavy reliance on engineers

➡️

Product teams analyze feedback to improve their products

😔

Overwhelmed by junk - difficult to spot useful feedback

😔

Lack the means to circle back to the users with updates

Configure

Collect

Consume

?

➡️

Users give feedback with various intents - express dissatisfaction, seek help, report issues

😔

Difficulties in articulating issues [you are here]

😔

Feedback often the only (ineffective) way to seek help

➡️

Users expect a response

😔

Instead they are left wondering what happened to their feedback - leading to distrust & apathy

Challenges & opportunities

Solutions

➡️

Users give feedback with various intents - express dissatisfaction, seek help, report issues

😔

Difficulties in articulating issues [you are here]

😔

Feedback often the only (ineffective) way to seek help

Collect

➡️

Users expect a response

😔

Instead they are left wondering what happened to their feedback - leading to distrust & apathy

?

Product teams analyze feedback to improve their products

➡️

Overwhelmed by junk - difficult to spot useful feedback

😔

😔

Lack means to circle back to the users with updates

Consume

Product teams set up feedback as per their requirements

➡️

Limited customizations

😔

Heavy reliance on engineers

😔

Configure

Challenges & opportunities

Solutions

➡️

Users give feedback with various intents - express dissatisfaction, seek help, report issues

😔

Difficulties in articulating issues [you are here]

😔

Feedback often the only (ineffective) way to seek help

Collect

➡️

Users expect a response

😔

Instead they are left wondering what happened to their feedback - leading to distrust & apathy

?

Product teams analyze feedback to improve their products

➡️

Overwhelmed by junk - difficult to spot useful feedback

😔

😔

Lack means to circle back to the users with updates

Consume

Product teams set up feedback as per their requirements

➡️

😔

Limited customizations

Heavy reliance on engineers

😔

Configure

Hypothesis

UX-PM collaboration

Research revealed that up to 40% of users in some NBU (Next Billion Users) countries are semi- to non-literate, and typing in local languages can be a major barrier—especially for NIU (Novice Internet Users) unfamiliar with standard keyboards.

UX and PM jointly hypothesized that introducing voice input could help these users share clearer, more useful feedback. Instead of jumping to solutions, I proposed a lightweight experiment using built-in speech-to-text features already available in most smartphone keyboards to quickly test this assumption.

“I always use voice because it’s faster and I don’t always know the spelling.” -user

“I always use voice because it’s faster and I don’t always know the spelling.” -user

Validation

The experiment outcomes were promising:

  • The experiment group exposed to the speech to text feature, showed a decline in volume of junk feedback reports by ~2% from 57,315 (56.86%) of 100,800 total volume to 40,548 (54.65%) of 74,192 total volume over the course of 1 month.

  • The experiment group submission rate improved marginally to 11.82% as compared to control group at 11.53%.

  • ~2% more feedback could be categorized using our pre-defined issue tags.

However, there was a caveat with this approach - the location of the default microphone button varied with the device manufacturers and their respective UIs. This reaffirmed the need to design a consistent experience.

Design brief

Design a dedicated voice typing feature that’s consistent across devices and easy to discover so that users can give us better feedback.

Guiding principles:

🔸

Familiarity & consistency across devices.

Familiarity & consistency across devices.

🔸

Ease of discovery.

🔸

Flexibility to switch between input types.

Flexibility to switch between input types.

Exploration 1 - user selects input type

➡️ Steps

  1. Select modality (type/speak).

  2. On selecting 'Speak', the mic is activated and the keyboard is replaced by the listening screen. The user can speak their issue and see it get transcribed in the text input field.

  3. The user can switch back to text mode.

  4. Make final edits.

  5. Submit the feedback.

✅ Pros

Easy to discover as it’s integrated into the flow of the conversation. #discoverability

Can adapt to user behavior over time. Based on past preferences, UI can default to either ‘Type’ or ‘Speak’.

⛔ Cons

Adds an extra step to the core flow.

User needs to switch between modalities to be able to make edits to the transcription.

Requires higher engineering effort.

Requires modifications to the core flow.

Exploration 2 - the morphing button

➡️ Steps

  1. Type right away or choose to speak.

  2. Describe the issue using voice, see it get transcribed in the text input field directly.

  3. Make edits to the transcript.

  4. Submit the feedback.

✅ Pros

Users can dictate and edit the transcript simultaneously without having to switch between modalities. #flexibility

Minimal distraction from the core flow. #familiarity

⛔ Cons

Adoption might be impacted as users will have the natural tendency to type when the keyboard is open.

Exploration 3 - the dictation screen

➡️ Steps

  1. Type right away or choose to speak.

  2. Describe the issue using voice in the dedicated dictation screen.

  3. Exit dictation screen.

  4. Edit the transcript.

  5. Submit the feedback.

✅ Pros

Dedicated dictation screen makes it easier to focus on the transcript with larger text.

⛔ Cons

Distracts from the core flow.

Ambiguity around the next step - will it submit the feedback directly or will the user have the option to edit?

A/B testing explorations 2 & 3 (qualitative)

UX-Research collaboration

Sample space

8-10 participants

Interview questions

Usability, comprehension, relevance and ease of use

Observations

Users preferred inline voice input (2) over full-screen (3) for its familiarity and ease of editing

A/B testing explorations 2 & 3 (quantitative)

UX-Engg. collaboration

Specifications for hand-off

Button states, error handling, animations

Metrics to measure success

Adoption, usability, engagement, quality, volume & submission rates

Launched & landed!

Exploration 2 - the morphing button emerged as the winner after running ~400,000 total sessions for each exploration (2 & 3).

What clicked?

  • Simple and straight forward to understand, navigate.

  • Editing on the same page feels intuitive.

  • Mic states are understood better.

  • For NBU markets, 41% reports were submitted using this option as compared to 26% for the other option (exploration 3, full-screen dictation).

  • For all markets, 36% reports were submitted using this option as opposed to 22% using the other option.

  • 1.64% higher adoption compared to the exploration 3.

This resulted in an average of 7% reduction in junk feedback volume over the course of 3 months after launch.

[Experience it live in the Google drive app on your Android device]

Thanks for the read!

A/B testing explorations 2 & 3 (qualitative)

UX-Research collaboration

Sample space

8-10 participants

Interview questions

Usability, comprehension, relevance and ease of use

Observations

Users preferred inline voice input (2) over full-screen (3) for its familiarity and ease of editing

A/B testing explorations 2 & 3 (quantitative)

UX-Engg. collaboration

Specifications for hand-off

Button states, error handling, animations

Metrics to measure success

Adoption, usability, engagement, quality, volume & submission rates

Launched & landed!

Exploration 2 - the morphing button emerged as the winner after running ~400,000 total sessions for each exploration (2 & 3).

What clicked?

  • Simple and straight forward to understand, navigate.

  • Editing on the same page feels intuitive.

  • Mic states are understood better.

  • For NBU markets, 41% reports were submitted using this option as compared to 26% for the other option (exploration 3, full-screen dictation).

  • For all markets, 36% reports were submitted using this option as opposed to 22% using the other option.

  • 1.64% higher adoption compared to the exploration 3.

This resulted in an average of 7% reduction in junk feedback volume over the course of 3 months after launch.

[Experience it live in the Google drive app on your Android device]

Thanks for the read!

A/B testing explorations 2 & 3 (qualitative)

UX-Research collaboration

Sample space

8-10 participants

Interview questions

Usability, comprehension, relevance and ease of use

Observations

Users preferred inline voice input (2) over full-screen (3) for its familiarity and ease of editing

A/B testing explorations 2 & 3 (quantitative)

UX-Engg. collaboration

Specifications for hand-off

Button states, error handling, animations

Metrics to measure success

Adoption, usability, engagement, quality, volume & submission rates

Launched & landed!

Exploration 2 - the morphing button emerged as the winner after running ~400,000 total sessions for each exploration (2 & 3).

What clicked?

  • Simple and straight forward to understand, navigate.

  • Editing on the same page feels intuitive.

  • Mic states are understood better.

  • For NBU markets, 41% reports were submitted using this option as compared to 26% for the other option (exploration 3, full-screen dictation).

  • For all markets, 36% reports were submitted using this option as opposed to 22% using the other option.

  • 1.64% higher adoption compared to the exploration 3.

This resulted in an average of 7% reduction in junk feedback volume over the course of 3 months after launch.

[Experience it live in the Google drive app on your Android device]

Thanks for the read!