Agent Building
Advanced
Always open

Agentic System for Global AI Product Competitive Intelligence

The rapid evolution of generative AI necessitates sophisticated competitive intelligence. This challenge tasks developers with building a multi-agent system using the OpenAI Agents SDK to autonomously monitor and analyze the global AI product landscape, inspired by headlines like ByteDance's Doubao 2.0 'agent era' upgrade. The system will leverage a team of specialized agents to gather data on new features, market positioning, and user sentiment for competing AI applications, synthesizing this information into actionable strategic insights.

Challenge brief

What you are building

The core problem, expected build, and operating context for this challenge.

The rapid evolution of generative AI necessitates sophisticated competitive intelligence. This challenge tasks developers with building a multi-agent system using the OpenAI Agents SDK to autonomously monitor and analyze the global AI product landscape, inspired by headlines like ByteDance's Doubao 2.0 'agent era' upgrade. The system will leverage a team of specialized agents to gather data on new features, market positioning, and user sentiment for competing AI applications, synthesizing this information into actionable strategic insights.

Datasets

Shared data for this challenge

Review public datasets and any private uploads tied to your build.

Loading datasets...
Evaluation rubric

How submissions are scored

These dimensions define what the evaluator checks, how much each dimension matters, and which criteria separate a passable run from a strong one.

Max Score: 5
Dimensions
5 scoring checks
Binary
5 pass or fail dimensions
Ordinal
0 scaled dimensions
Dimension 1accuratedataextraction

AccurateDataExtraction

Verifies that key factual data (features, dates, names) is correctly extracted and reported.

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 2insightfulanalysis

InsightfulAnalysis

Evaluates the depth and relevance of strategic implications and competitor comparisons.

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 3toolutilization

ToolUtilization

Checks if Portia AI and Ludwig were effectively integrated and used in the workflow.

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 4analysiscompleteness

AnalysisCompleteness

Percentage of required analysis fields correctly populated. • target: 90 • range: 0-100

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Dimension 5sentimentaccuracy

SentimentAccuracy

Accuracy of identified market sentiment compared to ground truth. • target: 0.85 • range: 0-1

binary
Weight: 1
Binary check

This dimension contributes its full weight only when the submission satisfies the requirement. Partial credit is not awarded.

Learning goals

What you should walk away with

Master the OpenAI Agents SDK for building role-based agent teams capable of dynamic tool use and multi-turn conversations.

Implement advanced function calling with Claude 4 Sonnet to enable agents to interact with external APIs and retrieve real-time data.

Design and integrate Portia AI as a specialized agent component for enhanced pattern recognition and anomaly detection within collected market data.

Leverage Ludwig for orchestrating the overall multi-step workflow, ensuring seamless transition between data collection, analysis, and reporting phases.

Utilize Prophet for real-time observability and tracing of agent actions, decisions, and output quality, facilitating debugging and optimization.

Develop a natural language interface using Hume AI to allow executive users to query the competitive intelligence system and receive concise, voice-synthesized summaries of findings.

Build a persistent memory mechanism for agents to retain context and learning across long-running analysis cycles, enhancing their analytical depth over time.

Start from your terminal
$npx -y @versalist/cli start agentic-system-for-global-ai-product-competitive-intelligence

[ok] Wrote CHALLENGE.md

[ok] Wrote .versalist.json

[ok] Wrote eval/examples.json

Requires VERSALIST_API_KEY. Works with any MCP-aware editor.

Docs
Manage API keys
Challenge at a glance
Host and timing
Vera

AI Research & Mentorship

Starts Available now
Evergreen challenge
Your progress

Participation status

You haven't started this challenge yet

Timeline and host

Operating window

Key dates and the organization behind this challenge.

Start date
Available now
Run mode
Evergreen challenge
Explore

Find another challenge

Jump to a random challenge when you want a fresh benchmark or a different problem space.

Useful when you want to pressure-test your workflow on a new dataset, new constraints, or a new evaluation rubric.

Tool Space Recipe

Draft
Evaluation
Rubric: 5 dimensions
·AccurateDataExtraction(1%)
·InsightfulAnalysis(1%)
·ToolUtilization(1%)
·AnalysisCompleteness(1%)
·SentimentAccuracy(1%)
Gold items: 2 (2 public)

Frequently Asked Questions about Agentic System for Global AI Product Competitive Intelligence