All case studies

Designing an Emotional AI Companion for Assisted Living

Designing an Emotional AI Companion for Assisted Living

From human-centered voice interaction to clinical monitoring and operational dashboards
From human-centered voice interaction to clinical monitoring and operational dashboards

Team

AiMA

Role

Senior Product Designer

Date

August 2025 - Now

Designing an AI for Elder Care in Assisted Living

AiMA was built for assisted living environments, where elderly users face social isolation and varying cognitive abilities — and where care teams need clear, actionable visibility without added operational burden.

This wasn’t just a tablet UI challenge. We had to design trust and emotional continuity for users aged 75+, while translating AI emotional signals into monitoring and alerts that staff could act on without alert fatigue.

AiMA is an ecosystem with three connected layers:

Emotional Interaction (Tablet Experience)

Clinical Monitoring & Alerts

Operations Dashboard (Tolkien)

This was the UI at New Relic before we started the "Entity Foundations" initiative.

Emotional Interaction Foundations

1. Context

AiMA was designed for elderly users in assisted living environments, many of whom experience social isolation, cognitive decline, and low digital literacy.

Traditional conversational interfaces are not built for users aged 75+. They often rely on complex UI patterns, fast-paced interactions, and abstract system feedback that can generate confusion or mistrust.

The goal was to design a calm, intuitive, and emotionally safe AI interaction that elderly users could adopt naturally — without feeling overwhelmed or intimidated.

2. Research

We conducted contextual research in assisted living environments, observing how elderly residents interacted with technology and how care staff supported them.

This included:

Observing real tablet usage behaviors

Identifying cognitive friction points

Understanding emotional triggers and trust signals

Mapping attention span and interaction rhythm

We also analyzed conversational design best practices and accessibility guidelines to ensure the experience would accommodate visual, auditory, and cognitive limitations.

3. Challenge

The main challenge was balancing emotional warmth with extreme simplicity.

We needed to:

Build trust in an AI system among vulnerable users

Reduce cognitive load to the minimum

Avoid visual noise and unnecessary controls

Provide clear system feedback (listening, thinking, speaking)

Ensure the interaction felt human but not deceptive

Additionally, the system had to work reliably in shared care environments, where interruptions and external noise were common.

4. Design Process

We defined a minimal interaction model based on three primary states:

Listening

Processing

Speaking

The interface was reduced to essential elements:

A central conversational area

Clear visual feedback for system state

A prominent “End Conversation” action

A simple text confirmation input

We iterated on visual density, typography scale, contrast, and spacing to ensure readability and calmness.

Prototypes were tested in real environments, and adjustments were made to pacing, button size, and system feedback clarity.

The result was an interaction model focused on emotional continuity rather than feature richness.

5. Collaboration with Other Teams

I worked closely with:

AI engineers to define system states and feedback timing

Care staff to validate usability in real scenarios

Developers to ensure interaction states were technically feasible

Stakeholders to align emotional goals with product strategy

Workshops and iterative reviews helped refine the balance between emotional design and operational constraints.

6. Results and Metrics

After deployment in pilot residences:

+33% increase in weekly interactions per user

75% of users returned to the system for consecutive sessions

+25% average session duration

66% of care staff reported improved perceived emotional engagement

The simplified interface reduced onboarding time and increased confidence among first-time users.

7. Final Conclusion

This project allowed New Relic to have a much more cohesive platform, where users could navigate more smoothly across different parts of the application. The established design language also created a solid foundation for future projects, improving team collaboration and accelerating development.

The hexagon, as a representation of the "entity" concept of a system, and the expression of its health status became a core part of the product.

These are some key screens in the product showing the new visual system and UI, even before the new branding was fully defined.

What needed to be improved:
Traditional conversational interfaces were not suitable for elderly users in assisted living environments.

Business Goals:
Increase emotional engagement. Build trust in AI. Improve user adoption.
Key Metrics:
Weekly Active Usage, Session Duration, Recurrence Rate, Staff Feedback.

What needed to be improved:
Traditional conversational interfaces were not suitable for elderly users in assisted living environments.

Business Goals:
Increase emotional engagement. Build trust in AI. Improve user adoption.

Key Metrics:
Weekly Active Usage, Session Duration, Recurrence Rate, Staff Feedback.

Tolkien Platform Foundations

1. Context

As AiMA scaled from pilot deployments to multiple assisted living sites, we needed a single platform to operate the service end-to-end.

Tolkien became the control center for:

Internal operations (devices, installations, sessions, incidents, QA)

Clinical visibility (alerts, resident status, trends)

Customer access (care managers and client teams) with role-based permissions

The goal was to build a system that could support day-to-day operations while also giving clients trustworthy, self-serve insight into their residents’ wellbeing and alert history.

2. Research

We mapped workflows across two distinct user groups:

Internal teams

  • Operations, support, QA, and product

  • Need: speed, debugging clarity, full access, auditability

Client teams

  • Care managers and supervisors

  • Need: clarity, trust, minimal noise, and privacy-safe views

We analyzed common failure points in monitoring tools:

  • Overly technical dashboards

  • Alert fatigue

  • Lack of context (why an alert happened)

  • Poor permission models that break trust

This research informed a dual-mode platform: operational depth for internal users + simplified clinical insight for clients.

3. Challenge

The biggest challenge was designing one platform for two audiences with conflicting needs:

Internal users needed full control and technical detail

Clients needed simplified, actionable insights

Both required trust, traceability, and privacy safeguards

We also had to ensure the platform scaled across multiple residences and deployments without becoming a maze of entities, filters, and exceptions.

4. Design Process

We structured Tolkien around a clear hierarchy:

Resident / Patient → Sessions → Signals → Alerts → Actions

Key design decisions:

A monitoring layer centered on sessions, quality score, and alert signals

Fast filtering and searching (by status, time range, quality indicators)

Alert states designed for actionability (open, acknowledged, resolved)

A “resident snapshot” pattern to summarize mood, vitals (when available), medication, and recent alerts

Role-based access patterns to expose the right level of detail to clients without operational noise

We iterated heavily on information density, scannability, and permission-aware UI states (what’s visible, what’s editable, what’s audit-logged).

5. Collaboration with Other Teams

I worked closely with:

Engineers to align entity models, permissions, and scalable UI architecture

Ops/support teams to validate real workflows and edge cases

Clinical/care stakeholders to validate alert usefulness and interpretation

Leadership to define what should be customer-facing vs internal-only

We reviewed iterations using real scenarios (a device failure, a low-mood alert, a conversation quality drop) to ensure the platform supported day-to-day decision making.

6. Results and Metrics

After adoption in production workflows:

  • -40% reduction in time to investigate incidents and alerts

  • +50% faster triage due to improved filtering and resident snapshots

  • 90% of client stakeholders could self-serve key insights without requesting internal support

  • Significant reduction in “false urgency” thanks to clearer alert resolution states

Tolkien became the shared source of truth for operational status and resident wellbeing monitoring.

7. Final Conclusion

Tolkien turned AiMA into an operable, scalable service — not just a conversational experience.

By designing a permission-aware platform that supports both internal operations and client visibility, we enabled trustworthy monitoring, faster response to risk signals, and a scalable foundation for future deployments.

What needed to be improved:
A growing AI care service required a scalable control center for operations and client-facing monitoring.
Business Goals:
Scale deployments, reduce operational workload, enable client self-serve insights, and increase trust in monitoring.
Key Metrics:
Time to Triage, Support Ticket Reduction, Client Self-Serve Rate, Alert Resolution Time.

What needed to be improved:
A growing AI care service required a scalable control center for operations and client-facing monitoring.

Business Goals:
Scale deployments, reduce operational workload, enable client self-serve insights, and increase trust in monitoring.

Key Metrics:
Time to Triage, Support Ticket Reduction, Client Self-Serve Rate, Alert Resolution Time.

AI-Augmented Product Delivery

1. Context

As AiMA evolved, we needed to accelerate how design decisions moved from Figma to production.

Traditional handoffs were slowing iteration cycles and introducing inconsistencies between design and implementation. At the same time, we were building both a marketing landing and product-facing interfaces that needed tight alignment with the system architecture.

The goal was to reduce friction between design and development while maintaining quality and consistency.

2. Research

We analyzed our existing workflow:

  • Static handoff files in Figma

  • Manual translation of components into code

  • Inconsistent token usage

  • Repeated UI implementation work

We explored emerging AI-assisted workflows and tooling that could connect design systems directly with codebases, including MCP-based integrations and AI copilots inside VS Code.

The opportunity was clear: reduce iteration time without compromising design intent.

3. Challenge

The main challenge was building a workflow that:

  • Preserved design system integrity

  • Reduced implementation time

  • Avoided “AI-generated chaos”

  • Kept developers in control of architecture

  • Enabled rapid UI iteration

We also needed to ensure that AI-assisted generation produced production-ready components, not throwaway prototypes.

4. Design Process

We implemented an AI-augmented workflow structured around:

1. Structured Design in Figma

  • Tokens for color, spacing, typography

  • Componentized UI patterns

  • Clear hierarchy and naming

2. MCP Integration

  • Direct linkage between Figma and repository structure

  • Component-level alignment

3. Codex in VS Code

  • Assisted generation of UI components

  • Rapid iteration on layout and responsiveness

  • Reduced boilerplate

4. Iterative Validation

  • Manual review and refinement

  • Alignment with system architecture

  • E2E validation when needed

This workflow was used to build the new AiMA landing and accelerate internal UI iteration.

5. Collaboration with Other Teams

I worked closely with developers to:

  • Define how tokens map to code

  • Review generated components

  • Maintain architectural integrity

  • Establish guardrails for AI-assisted generation

Instead of replacing engineering effort, the system augmented it — reducing repetitive implementation work while preserving ownership.

6. Results and Metrics

-30% reduction in UI implementation time

  • Significant decrease in visual inconsistencies

  • Faster landing page iteration cycles

  • Improved alignment between design system and production components

The workflow also reduced dependency bottlenecks between design and development.

7. Final Conclusion

This AI-augmented delivery model allowed us to move from traditional handoff to collaborative iteration.

By connecting structured design systems with AI-assisted development, we accelerated product evolution while maintaining control and quality.

This approach reflects a broader shift in how modern product teams can operate at speed — without sacrificing design integrity.

What needed to be improved:
Traditional design-to-development handoffs were slowing iteration and introducing inconsistencies.
Business Goals:
Increase delivery speed. Improve design-code alignment. Reduce redundant implementation work.
Key Metrics:
Implementation Time, UI Consistency Rate, Iteration Speed, Dev-Design Alignment..

What needed to be improved:
Traditional design-to-development handoffs were slowing iteration and introducing inconsistencies.

Business Goals:
Increase delivery speed. Improve design-code alignment. Reduce redundant implementation work.

Key Metrics:
Implementation Time, UI Consistency Rate, Iteration Speed, Dev-Design Alignment..

Let's

Talk

I would love to help you turn complex visions
into user-centered solutions that matter.

Let's

Talk

I would love to help you turn complex visions into user-centered solutions that matter.

Let's

Talk

I would love to help you turn complex visions
into user-centered solutions that matter.

Create a free website with Framer, the website builder loved by startups, designers and agencies.