Agent Building
Advanced
Always open

Build a Hyper-Personalized Voice Assistant Agent

This challenge tasks developers with creating an advanced, real-time voice assistant designed for hyper-personalization and proactive device or application management. The solution must leverage Mastra AI's agentic workflows and robust memory capabilities to maintain long-term user preferences and context. Claude Opus 4.6 will be integrated for its nuanced conversational understanding and empathetic response generation, while ElevenLabs will provide natural, low-latency speech synthesis and recognition for a seamless voice user experience. The assistant should dynamically adapt its responses and actions based on the user's historical interactions and real-time device state. Participants will focus on designing reactive agent workflows in Mastra AI, implementing bidirectional real-time speech interaction, and developing custom tools for device integration. The ultimate goal is to deliver a highly intuitive and personalized voice interface that anticipates user needs and acts intelligently, enhancing the user's interaction with their digital environment.

Challenge brief

What you are building

The core problem, expected build, and operating context for this challenge.

This challenge tasks developers with creating an advanced, real-time voice assistant designed for hyper-personalization and proactive device or application management. The solution must leverage Mastra AI's agentic workflows and robust memory capabilities to maintain long-term user preferences and context. Claude Opus 4.6 will be integrated for its nuanced conversational understanding and empathetic response generation, while ElevenLabs will provide natural, low-latency speech synthesis and recognition for a seamless voice user experience. The assistant should dynamically adapt its responses and actions based on the user's historical interactions and real-time device state. Participants will focus on designing reactive agent workflows in Mastra AI, implementing bidirectional real-time speech interaction, and developing custom tools for device integration. The ultimate goal is to deliver a highly intuitive and personalized voice interface that anticipates user needs and acts intelligently, enhancing the user's interaction with their digital environment.

Datasets

Shared data for this challenge

Review public datasets and any private uploads tied to your build.

Loading datasets...
Evaluation rubric

How submissions are scored

These dimensions define what the evaluator checks, how much each dimension matters, and which criteria separate a passable run from a strong one.

Max Score: 6
Dimensions
6 scoring checks
Binary
6 pass or fail dimensions
Ordinal
0 scaled dimensions
Dimension 1task_completion_accuracy

task_completion_accuracy

Checks if all requested tasks were completed correctly and without errors based on the user's voice commands.

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 2memory_recall_effectiveness

memory_recall_effectiveness

Verifies if personalized information from long-term memory was correctly accessed and utilized in responses or actions.

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 3real_time_responsiveness

real_time_responsiveness

Assesses if responses were generated and spoken within acceptable latency for a natural voice interaction (e.g., < 1 second).

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 4personalization_score

personalization_score

Degree to which the assistant adapted its behavior and responses based on user preferences and history. • target: 0.85 • range: 0.6-1

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 5conversational_turn_accuracy

conversational_turn_accuracy

Percentage of conversational turns correctly understood, interpreted, and responded to by the assistant. • target: 0.92 • range: 0.8-1

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 6average_response_latency_ms

average_response_latency_ms

Average time taken (in milliseconds) from user speaking to assistant starting to respond. • target: 300 • range: 50-1000

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Learning goals

What you should walk away with

Master building reactive agent workflows with Mastra AI, utilizing its state management and tool orchestration for dynamic task execution.

Implement real-time bidirectional speech interaction using ElevenLabs Text-to-Speech and Speech-to-Text APIs for a seamless voice user experience.

Design and integrate a long-term memory system within Mastra AI to capture and recall user preferences, historical interactions, and device context for hyper-personalization.

Leverage Claude Opus 4.6's advanced conversational capabilities for nuanced natural language understanding, intent recognition, and empathetic response generation.

Develop custom tools for the Mastra AI agent to interact with simulated device settings (e.g., calendar, reminders, app control) and external APIs.

Integrate the voice assistant with Ellipsis to provide a rich conversational UI alongside the voice interface, enhancing accessibility and interaction modes.

Implement robust error handling and conversational recovery strategies within the Mastra AI agent to maintain fluidity during complex interactions or misunderstandings.

Start from your terminal
$npx -y @versalist/cli start build-a-hyper-personalized-voice-assistant-agent

[ok] Wrote CHALLENGE.md

[ok] Wrote .versalist.json

[ok] Wrote eval/examples.json

Requires VERSALIST_API_KEY. Works with any MCP-aware editor.

Docs
Manage API keys
Challenge at a glance
Host and timing
Vera

AI Research & Mentorship

Starts Available now
Evergreen challenge
Your progress

Participation status

You haven't started this challenge yet

Timeline and host

Operating window

Key dates and the organization behind this challenge.

Start date
Available now
Run mode
Evergreen challenge
Explore

Find another challenge

Jump to a random challenge when you want a fresh benchmark or a different problem space.

Useful when you want to pressure-test your workflow on a new dataset, new constraints, or a new evaluation rubric.

Tool Space Recipe

Draft
Evaluation
Rubric: 6 dimensions
·task_completion_accuracy(1%)
·memory_recall_effectiveness(1%)
·real_time_responsiveness(1%)
·personalization_score(1%)
·conversational_turn_accuracy(1%)
·average_response_latency_ms(1%)
Gold items: 2 (2 public)

Frequently Asked Questions about Build a Hyper-Personalized Voice Assistant Agent