Multi-modal for car experience


Role

Conversation Designer

Multi-Modal UX Designer

UI Designer


Deliverables

Multi-modal UX design

  • Low- fidelity wireframes with multi-modal design

UI Design

  • High-fidelity mockups and prototypes

  • UI elements

Conversation Design

  • Competitive analysis

  • User research

  • User journeys and task flows

  • Sample dialogues and error handling solutions

  • Intent design and prompt design

  • Usability tests and findings


Project Specifications

Duration

  • 8 WEEKS

Company

  • NUANCE (Cerence)

Tools

  • Omnigraffle

  • Sketch


Introduction

At the heart of Nuance's ambition to redefine the car experience with AI-powered solutions, I embarked on a journey to integrate eye-gaze detection into Dragon Drive, now celebrated as Cerence Drive. This narrative unfolds over eight transformative weeks, where I navigated through the roles of Conversation Designer, Multi-Modal UX Designer, and UI Designer, crafting a multi-modal UX design that bridges the car with the world outside.

The Challenge

Nuance envisioned a future where the car is not just a vehicle but a hub of connectivity. The mission was clear: to design features showcasing this vision through visuals, eye gaze, and voice. The challenge was to not only envision but also execute a multi-modal in-car experience that could seamlessly integrate with the driver's environment.

Research: unveiling the potential

My journey began with a deep dive into existing research on the Nuance Dragon Drive head unit. By studying usage frequency and feature-scenario connections, I gleaned insights into the features that drivers value in their interactions with their cars.

Brainstorming: Imagining the Possibilities

Envisioning a user immersed in the driving experience, I asked, "What does the user see, and what information do they seek?" This led to the conception of GAZE functionality, focusing on both the interior and exterior of the car. Sketching these scenarios, I explored the connectivity between the car and the external world, notably through restaurants, hotels, shopping malls, and general queries.

 
What Can Users See?

What Can Users See?

Objects outside a car

Concept validation

After considering the available data sources, existing voice supported domains, user pain points, popular scenarios, and optimal outcomes, I narrowed the ideas down to a few topics:

  • restaurants

  • hotels

  • general object query

  • shopping malls

After this, I set out to explore the restaurant topic in detail through use case mapping.

Designing the Experience

From Use Case Mapping to Storyboarding

Collaboration with PMs, developers, and speech engineers helped refine and validate the feasibility of our ideas. Storyboarding became a pivotal tool in communicating UX ideas, receiving stakeholder feedback on technology constraints and feasibility.

Use Case Map - Restaurants

Use Case Map - Restaurants

gaze_story.jpg

VUI Design: Crafting Conversations

I initiated the VUI design with a focus on the "happy path," progressively integrating complexity. This approach ensured a natural flow, mimicking human conversation while collaborating closely with engineers to align on system logic and database parameters.

gaze_happypath.jpg

A Happy Path Example

VUI elements

To complete dialog flow, I usually carefully think through all the intents, slots, utterances, prompts, GUI screens, earcon, hints, and system logic. During this phase, I worked very closely with engineers to make sure I have all the parameters that I needed in the database, and to make sure that the system logic is feasible.

gaze_happypath_analyze.jpg

GUI Design: Enhancing Interaction

Acknowledging the voice-first nature of in-car features, I aimed to augment this with an intuitive graphical user interface. Starting with low-fidelity wireframes, I engaged stakeholders in validating the design's feasibility, eventually evolving these into high-fidelity mockups. This phase underscored the importance of simplicity and clarity in in-car screens, where driver attention is paramount.

Wireframe

Wireframe

High Fidelity Mockup

High Fidelity Mockup

My full set of deliverables for this project includes: VUI flow, GUI wireframes & hi-fi mockups, prompt lists, and intent lists. After finishing all the designs, I moved to the testing phase.

Testing and Refinement

A week of intensive usability testing across three offices brought invaluable insights, leading to several refinements in our design. This phase was crucial for validating the user experience and ensuring the design met our users' needs and expectations.

AI Training

We constantly collected possible utterances/phrases to deliver to the development team, so that they could better organize the data and train the AI.

Results and Reflections

The feature's launch marked a significant milestone, with its popularity among users affirming the project's success. Nuance, now Cerence, not only secured new clients but also featured this innovation prominently on their website.

Reflecting on this journey, I recognize the evolving nature of multi-modal design and its expanding technological landscape. This project underscored the importance of considering multiple dimensions of user interaction, from visual and auditory to future possibilities of sensory engagement.

Final Thoughts

This project was a testament to the power of collaboration, innovation, and user-centered design. It taught me the intricacies of multi-modal design and the need for a holistic approach to UX, blending touch, gaze, voice, and potentially even more senses in the future.