Evaluation and Benchmarking of Generative and Agentic AI Systems: A Comprehensive Survey
DOI:
https://doi.org/10.31224/6009Keywords:
Agentic AI, Generative AI, AI Evaluation and Benchmarking, Autonomous Intelligent Systems, Tool-Augmented Language Models, Multi-Agent Systems, Long-Horizon Planning, Memory and Reasoning Evaluation, Adaptive Monitoring, AI Safety and Robustness, Multimodal AI EvaluationAbstract
The rapid emergence of generative and agentic artificial intelligence (AI) has outpaced traditional evaluation practices. While large language models excel on static language benchmarks, real-world deployment demands more than accuracy on curated tasks. Agentic systems use planning, tool invocation, memory and multi-agent collaboration to perform complex workflows. Enterprise adoption therefore hinges on holistic assessments that include cost, latency, reliability, safety and multi-agent coordination. This survey provides a comprehensive taxonomy of evaluation dimensions, reviews existing benchmarks for generative and agentic systems, identifies gaps between laboratory tests and production requirements, and proposes future directions for more realistic, multi-dimensional benchmarking.
Downloads
Downloads
Posted
License
Copyright (c) 2025 Manish Shukla

This work is licensed under a Creative Commons Attribution 4.0 International License.