Back to Prompt Library
implementation

Implement Gemini Multimodal Interaction

Inspect the original prompt language first, then copy or adapt it once you know how it fits your workflow.

Linked challenge: Gemini-powered Voice Navigator Agent

Format
Text-first
Lines
1
Sections
1
Linked challenge
Gemini-powered Voice Navigator Agent

Prompt source

Original prompt text with formatting preserved for inspection.

1 lines
1 sections
No variables
0 checklist items
Extend your Google ADK agent to use Gemini 1.5 Pro for processing user utterances and generating responses. Your agent should be able to interpret natural language commands related to navigation and safety. Show how you would configure the ADK to use Gemini as the primary LLM for reasoning and how to pass a user's voice input to Gemini and then use Gemini's response to drive the `speak` tool. Also, outline how you would pass environmental context (like current location and activity) to Gemini.

Adaptation plan

Keep the source stable, then change the prompt in a predictable order so the next run is easier to evaluate.

Keep stable

Hold the task contract and output shape stable so generated implementations remain comparable.

Tune next

Update libraries, interfaces, and environment assumptions to match the stack you actually run.

Verify after

Test failure handling, edge cases, and any code paths that depend on hidden context or secrets.