Prompt Content
Design an evaluation pipeline using Evidently AI to assess the 'ThreatDetectionAndClassification' task. Describe how you would collect the agent's outputs and ground truth, and define key metrics Evidently AI should track, such as 'ThreatTypeAccuracy' and 'SeverityClassificationF1Score'. Provide a Python snippet demonstrating how to initialize an Evidently AI monitoring dashboard and log relevant data from your agent's performance.
```python
import evidently
from evidently.report import Report
from evidently.metric_preset import DataDriftPreset
from evidently.metric_preset import TextOverviewPreset # For LLM specific outputs
def evaluate_threat_detection(actual_outputs, ground_truth):
# Prepare data for Evidently AI
# ... e.g., combine actual_outputs and ground_truth into a DataFrame ...
report = Report(metrics=[
DataDriftPreset(), # For input/output data drift
# Add custom metrics for classification accuracy, F1 score etc.
# Evidently AI might require custom metric definitions for direct LLM eval,
# or conversion of LLM outputs to structured data first.
])
# report.run(reference_data=ref_df, current_data=current_df)
# report.save_html("threat_detection_report.html")
# Conceptual usage:
# agent_output = threat_detection_agent(sample_input)
# expected_output = get_ground_truth(sample_input)
# evaluate_threat_detection([agent_output], [expected_output])
```Try this prompt
Open the workspace to execute this prompt with free credits, or use your own API keys for unlimited usage.
Related Prompts
Explore similar prompts from our community
Usage Tips
Copy the prompt and paste it into your preferred AI tool (Claude, ChatGPT, Gemini)
Customize placeholder values with your specific requirements and context
For best results, provide clear examples and test different variations