This project is protected

Enter the password to view this case study.

That's not it — try again.
Case Study — PhotoG

I slowed the AI down on purpose.

Four AI agents, one marketing platform — and most users left before publishing a single post. I redesigned the system's trust architecture, introducing intentional friction, progressive disclosure, and human-in-the-loop gates that turned autonomous AI into a collaborative tool.

Role
Product Designer
Focus
Agentic UX & Trust Architecture
Team
2 Designers
Timeline
8 Weeks
app.photog.ai
PhotoG welcome screen with four AI agent characters
0.1
Context

An AI marketing team in one product

PhotoG is an agentic AI marketing platform for e-commerce brands. Users enter their product details, and four AI agents collaborate to deliver everything from industry analysis to a published social media post. A full marketing department, automated.

The product existed before I joined. It worked — technically. The agents could research, strategize, create, and publish. But most users left before publishing a single post.

Research Analyst
Industry data & competitor analysis
Brand Strategist
Positioning & reference posts
Creative Director
Post content & product images
Operations Manager
Scheduling & publishing
0.2
Problem

Four problems, one pattern.

I ran moderated sessions with beta users. The existing interface dumped everything — research, strategy, content — into one flat page. Four agents worked behind the scenes, but users couldn't tell who was doing what, where they were in the process, or how to give the AI better input.

01

Flat information dump

Research reports, market data, and strategy all rendered on one long scrolling page. No visual hierarchy between what four different agents produced.

02

Invisible agents

Four AI agents worked behind the scenes, but the interface never showed who was doing what. Users saw output appear with no sense of process or progress.

03

No user control

The system moved from research to strategy to content automatically. No checkpoints, no approval gates. Users couldn't intervene or steer the output.

04

Users couldn't articulate what they wanted

Most users were small e-commerce owners, not marketers. They didn't know how to describe their brand positioning or target audience. The system expected clear input but gave no help forming it.

app.photog.ai
PhotoG's original interface — flat document page with no agent visibility or user control
1.0
Strategy

I made the AI slower on purpose

Before PhotoG, I worked at an advertising agency designing marketing workflows for real clients. That experience taught me something: when clients pay for creative work, they want to be part of the process. They want to see options, give feedback, and feel ownership over the result. AI shouldn't be different — if users feel like passengers watching a machine make decisions for them, they leave.

I spent the first two weeks embedded with the engineering team, learning how each agent worked under the hood — what it could predict, where it hallucinated, how it handed off context between stages. Understanding the model's behavior shaped every design decision that followed.

Then I designed friction back into the system — deliberate pauses where users could review, edit, and decide before the next agent took over.

01

Flow Audit

Mapped every step of the existing product across all five departments — user input, marketing, brand, design, and operations. Documented where users lost context, where the system auto-advanced without consent, and where output had no structure.

02

Human-in-the-Loop Gates

A confirmation step between every agent. Each gate shows what the agent produced, lets users edit it, and asks "Good?" before the next agent starts. Users can loop back at any point.

03

Guided Input & Templates

Most users weren't marketers — they didn't know how to describe brand positioning. I designed templates and structured prompts that helped users express what they wanted, so the AI had better input to work with.

04

Dual Panel with Chat

Left: ongoing conversation where users can ask questions and steer the AI at any point. Right: the editable deliverable. Users can intervene through chat, not just through gates.

Before
Original Pipeline
The entire flow ran left-to-right with no user checkpoints. Every arrow was automatic.
Submit
Product images
Product info
Social accounts
Brand Data Analysis
Industry — trends, SWOT, market size
Customer — segments, positioning
Product — features, pricing, audience
Competitor Intel
Competitor SOV
Keyword mining
Content planning
Account Matrix
Official accounts
KOL / KOC selection
Content keywords per channel
Content Creation
Select template
Generate copy
Generate images & video
Publish
Schedule timing
Cross-platform sync
Performance tracking
Each stage wrote to a separate database — no shared context between agents
No user gates between stages
No review before publish
Siloed databases
Users couldn't track progress
After
Redesigned Pipeline
Every stage has a confirmation gate. Users review, edit, and approve before the next agent starts. Chat is always accessible.
User Input
Templates & guided prompts
Brand details
Product images
Research Analyst
Data visualization
Industry analysis
Competitor report
Good?
Edit · Chat
Brand Strategist
Positioning
Tone guidelines
Reference posts
Good?
Edit · Chat
Creative Director
Post copy
Product images
Video content
Send?
Preview · Edit
Publish
Schedule timing
Cross-platform
Performance tracking
Chat accessible at every stage — users can ask questions, steer output, and loop back at any point
Users can loop back to any previous stage from any gate
Confirmation gate at every stage
Chat always accessible
Guided input for users
Loop-back from any gate
2.0
Solution

Four agents, three gates, one story

The redesigned flow treats each agent as a chapter. You don't move to the next until you've reviewed what the current one produced — and made a conscious decision to proceed.

app.photog.ai — Research Analyst
Stage 1 — Research Analyst
Design Decision 01

Intent previews as autonomy boundaries

The intent preview shows what the next agent plans to do, what it needs from you, and what happens if you say no. Every transition requires a deliberate click — because the person needs a locus of control.

app.photog.ai
Intent preview pattern
Design Decision 02

Left side reasons. Right side proves.

The conversational panel follows the AI's reasoning — the "why." The deliverable panel shows the output — the "what." Editable, downloadable, real.

app.photog.ai
Dual panel layout
3.0
Evolution

The first version wasn't right either.

My first iteration introduced the dual-panel concept and agent stepper — but usability sessions showed it still wasn't landing. So I iterated.

First Iteration

Right structure, missing control.

Introduced the dual-panel layout and agent visibility. But the flow still auto-advanced between agents — users could see what was happening, but couldn't steer it.

Final Version

Three gates. Full user control.

Added "Continue" gates between every agent transition. Users review each deliverable and make a conscious decision to proceed. The stepper, dual-panel, and gates work together as a trust system.

First Iteration
app.photog.ai
First iteration — dual-panel layout with no confirmation gates
Final Version
app.photog.ai
Final version — full user control with confirmation gates between every agent
4.0
Designing for Failure

AI output is never final. Users are.

LLMs don't produce perfect results — they produce starting points. The old design treated AI output as a finished deliverable. The redesign treats every output as a draft that the user shapes into something they trust. Two correction paths run through the entire product: direct editing and natural language.

Direct Editing

Every deliverable panel is fully editable. Users can rewrite copy, adjust data points, remove sections, or restructure the report — directly in the document, the same way they'd work in any editor. The AI produces the first draft; the user owns the final version.

Natural Language Adjustment

The chat panel is always open alongside the deliverable. Users can tell the agent to change tone, focus on a different competitor, shorten a section, or regenerate with new constraints — without leaving the current stage. The agent revises in place.

Review Before Advancing

The confirmation gate at each stage isn't just a "next" button. It's the moment where users decide: is this good enough to build on? If not, they edit, chat, or loop back. Nothing moves forward without a conscious decision.

Loop Back Anytime

If a later stage reveals a problem with earlier work — the strategy doesn't fit, the research missed a competitor — users can return to any previous stage, make changes, and let the downstream agents rebuild from the corrected foundation.

app.photog.ai — Editing post content
Creative Director stage showing editable title, caption, and hashtag fields — users can directly modify AI-generated content

The Creative Director stage: AI generates the post copy, images, and hashtags — but every field is editable. Users rewrite the title, adjust the caption, swap images, or ask the agent to regenerate via chat on the left.

5.0
Leadership

What I owned, and how we worked.

Two-person design collaboration. I owned the interaction architecture, established the design principles, and made the strategic case to stakeholders that slowing down the AI was the right call.

Interaction Architecture

Defined the system model — stage-gate pattern, dual-panel layout, progressive disclosure hierarchy, and the confidence visualization framework. Worked with engineering on real-time agent state changes, WebSocket thinking indicators, and session persistence.

Design Principles

Three principles governed every decision: "Demonstrated value before delegation," "Autonomy boundaries at every transition," and "Confidence over speed." The hardest sell was convincing stakeholders that adding friction to an AI product was a feature. We shipped the full redesign in 8 weeks.

AI Character Design

Designed the 3D character avatars for each AI agent. Concepted the personalities, art-directed the visual style, and generated the final characters using AI image tools. This bridged product design and brand — giving abstract agents a human-scale identity.

Four 3D AI agent characters — Research Analyst, Brand Strategist, Creative Director, and Operations Manager
6.0
Impact

Slower AI. Better numbers.

Significant
Reduction in drop-off before first publish
Internal beta — users who previously abandoned now completed the full flow
In control
Users reported feeling in control of the AI for the first time
Post-session feedback — a complete reversal from the old design where no one felt they could steer the system
Funded
Stakeholders approved v4.0 development cycle
Based on beta results and the agentic design framework we presented

After launch, the platform scaled to production. Six months in:

100K+
Users across 200+ countries
Platform adoption in the first six months post-launch
−75%
Content production cost
Compared to traditional marketing workflows for equivalent output
+60%
User engagement
Average increase in content interaction rates for active accounts
7.0
Reflection

What I'll carry forward

Friction is a design material

The most impactful thing I designed was a pause — a moment where someone stops and thinks before continuing.

Trust is built through showing, not telling

Users preferred watching the AI work step-by-step over getting instant results. Trust is built through accumulation, not assertion.

Design systems are now AI contracts

The confidence indicator, intent preview, and autonomy gate aren't UI components — they're contracts that define how humans and AI collaborate.

The designer's job is changing

The hardest problems in AI aren't about making models smarter — they're about making smart models legible, controllable, and worthy of trust. That's a design problem.

← Back to home Next: LastMinute →
© 2026 Han Meng — Product Designer