Context and my role
A zero-to-one AI product during a company pivot.
I worked as the sole Product Designer on a small cross-functional team (PM, x2 engineers) defining Let’s Do This’s first AI support system for event organisers.
At the time the company was doubling down on AI-powered tooling for event organisers. Rather than relying on off-the-shelf chatbots, we built Messenger natively so it could integrate with event data and organiser workflows.
My role spanned discovery with organisers, shaping the product strategy, designing the organiser tooling and customer conversational interface, and working closely with engineering to ship the first production release.

In this case study
This project involved designing a system that organisers trust to communicate with their customers.
This case study explores:
Designing an AI product organisers trust to communicate with their customers.
Turning complex operational workflows into simple tools.
Creating a confidence system to make AI reliability visible.
Designing operator tooling for event organisers.
Crafting conversation patterns, interaction states and visual systems
Connecting design decisions to product strategy and measurable outcomes.
What success looked like
Before designing Messenger, we defined how we would measure success.
Resolution Rate
The percentage of questions the agent can handle itself.
CSAT
Whether participants are happy with the responses.
Time to Launch
Whether organisers can get live without help.
Adoption
Installs, organisers going live, and whether the sales team demos it.
Our initial goal was roughly 60% autonomous resolution, which would represent a meaningful reduction in manual support workload.
Understanding the problem
Organisers consistently described the same pattern: inboxes filled with repetitive participant questions about transfers, refunds, start times and logistics.
Support volume spikes in the weeks before race day, leaving organisers answering the same questions instead of focusing on delivering the event.

The real barrier wasn’t AI. It was trust.
Organisers weren’t worried about setting up Messenger. They were worried about what would happen if it got something wrong. Turning the assistant on meant letting software speak on behalf of their event.
That reframed the challenge: adoption would depend less on AI capability and more on organiser confidence.
Understanding the forces behind adoption
To understand whether organisers would adopt an AI support agent, I used the Jobs to be Done forces framework. Speaking to several organisers made the problem very obvious. Their inboxes are full of the same questions, but hesitant to rely on AI without understanding how it behaves.

This highlighted four forces shaping the decision:
Push → "I'm drowning in the same questions every week."
Pull → "If this works, it could save us loads of time."
Anxieties → "What if it gives runners the wrong information?"
Inertia → "We've always just handled support ourselves."
Designing Messenger therefore meant doing more than building an AI assistant. We needed to reduce anxiety, build confidence, and make the value obvious quickly.
Reducing anxiety, building confidence and creating excitement
Introducing an AI system that would speak directly to participants created a new kind of risk for organisers. Before installing Messenger, they needed to feel confident that the system understood their event and would respond safely.
Designing Messenger therefore meant doing more than building an assistant.
We needed to reduce anxiety, make the system’s behaviour visible, and quickly show the value it could provide.
Instant comprehension and excitement
The Overview screen is the first thing organisers see.
Rather than introducing Messenger as an AI product, we framed it around a moment organisers immediately recognised:
“Imagine race week without inbox overload.”
This reframes Messenger around the outcome organisers care about rather than the technology behind it.
A short video demonstrates the assistant answering real participant questions, helping organisers immediately understand what the product does and why it matters. Alongside this, a guided checklist breaks the setup process into clear steps, turning an unfamiliar AI system into something approachable and manageable.
Within seconds organisers should understand the value of Messenger, see how it works, and feel confident enough to keep going. Huge props to Luke Willetts for the product marketing video.
Starting small to make adoption feel safe
Discovery conversations revealed that organisers rarely deploy new tools across all events immediately. Instead, they prefer to test changes on a single event first.
Messenger therefore encourages organisers to start by syncing one event page. This creates a bridge between the organiser’s website and the assistant, allowing Messenger to learn from the event’s existing content.
Using the event website as the source of truth also reduces a key anxiety organisers raised: the risk of information becoming outdated. Messenger automatically re-syncs this content over time, ensuring responses remain aligned with the organiser’s site.
Making AI reliability visible
Even with content synced, organisers still needed confidence that Messenger could answer participant questions correctly.
To make reliability visible, Messenger tests a set of common participant questions against the organiser’s content before launch.
Through organiser interviews, support team insights, and AI analysis of event websites, we identified 27 questions participants consistently ask. Rather than trying to cover every scenario, the goal was to cover enough real-world cases that organisers could feel confident installing Messenger.
To reduce overwhelm, the questions were grouped into three tiers:
• Launch-ready — high-risk questions involving bookings or money
• Support reducers — repetitive questions that dominate inbox volume
• Confidence builders — useful but less critical participant queries
Each question was broken down into 6–10 criteria that determine whether Messenger can answer safely. For example, answering “Can I get a refund?” requires clear information about eligibility, deadlines, fees and the refund process. If key signals are missing, Messenger flags the answer as uncertain and recommends improving the content.

Installing Messenger was the moment of real trust
Up to this point, organisers are exploring Messenger inside the product. Installing it means leaving the platform, sending a code snippet to their website team, and trusting that the assistant will now speak directly to participants.
If this moment feels risky or confusing, adoption stops.
The installation flow was therefore designed to feel simple, safe and reversible. Clear instructions guide organisers through adding Messenger to their site, while reassuring them that the assistant can easily be disabled or removed if needed.
This framing helped reduce the perceived risk of trying Messenger for the first time.

Designing the Monday morning view
Once Messenger was live, organisers needed a simple way to understand how it was performing and where to focus their attention.
To design this view, I ran several rounds of card sorting, explored the opportunity space with AI, and tested early concepts with organisers.
One question shaped the entire structure:
“It’s Monday morning after race weekend. You open your laptop with a cup of tea. What do you want to see first?”

The page hierarchy reflects the answer organisers consistently gave:
Top opportunities to reduce escalations → Messenger highlights where improving content could increase automation and prevent participant questions from reaching support.
Support activity forecast → Upcoming spikes in participant questions are predicted based on event timelines so organisers can prepare content ahead of time.
Performance overview → Key metrics reassure organisers that the system is operating as expected.
Suggested improvements → Each recommendation shows the potential impact, such as estimated automation increases or how many questions it could prevent.
What participants are asking → Real participant questions reveal the issues customers are currently trying to solve.
Messenger doesn’t just report what happened. It highlights where organisers can improve coverage before the next race week begins.
Designing the conversational interface
While organisers configure and operate Messenger, participants interact with it through a conversational interface embedded directly on event websites.
Designing this experience required more than simple chat bubbles. The system needed to handle multiple input methods, structured data, feedback states, and different event brands while remaining clear and trustworthy.

Clear input and feedback
Participants interact with Messenger through both text and voice. Designing the input system meant clearly communicating what the interface was doing at each stage of interaction.
We defined states for typing, voice recording, transcription and message confirmation so participants always understand how their input is being captured and processed.

Capturing structured data safely
Some conversations require participants to provide structured information such as email addresses, booking details or registration information.
Instead of free-form chat, Messenger uses structured inputs within the conversation. This allows the system to validate information in consistent formats while reassuring participants that sensitive data is handled securely rather than appearing as plain text in a chat thread.

Designing for many brands
Messenger needed to feel native across a wide range of event websites. We developed a flexible colour system that adapts to organiser branding while preserving contrast, readability and accessibility across light and dark surfaces.
Together these interaction patterns ensure Messenger feels reliable and predictable for participants, reinforcing the trust organisers need to confidently deploy the assistant.
From support tool to communication platform
Messenger began as a way to automate repetitive participant support questions. But as organisers started interacting with the system, a broader opportunity became clear.
Participants naturally prefer conversational interfaces over searching through event websites. Instead of static FAQs and one-way emails, Messenger created the foundation for ongoing conversations between organisers and participants.
Starting with chat allowed us to prove the value safely, but the long-term vision was broader: one AI assistant helping organisers communicate with participants across multiple channels including chat, email and messaging platforms.
This work established the foundation for that future.
Designing AI products means designing trust
The biggest lesson from this project was that AI capability alone doesn’t drive adoption.
Organisers weren’t evaluating Messenger based on how sophisticated the model was. They were evaluating whether they could trust the system to represent their event.
Designing Messenger therefore meant focusing less on the AI itself and more on the surrounding system: making reliability visible, reducing anxiety during setup, and giving organisers clear control over how the assistant behaves.
Confidence is the product. AI is just the mechanism.