Evaluate Summary and Entity Extraction

testingChallenge

Prompt Content

Run the provided `DocumentSummarization` and `EntityExtractionAndKnowledgeGraphQuery` evaluation tasks against your implemented system. Analyze the results reported by LangFuse and the evaluation harness. Identify areas for improvement in your RAG strategy, chunking, or prompt engineering for Gemini 2.5 Pro. Describe how you would iterate on your solution to improve factual accuracy and query recall based on the evaluation metrics.

Try this prompt

Open the workspace to execute this prompt with free credits, or use your own API keys for unlimited usage.

Usage Tips

Copy the prompt and paste it into your preferred AI tool (Claude, ChatGPT, Gemini)

Customize placeholder values with your specific requirements and context

For best results, provide clear examples and test different variations