As complex AI agents become common, standard evaluation isn't enough. This presentation provides a structured overview of the critical field of agentic system evaluation. We will briefly explore common single and multi-agent patterns, delve into the fundamental reasons why rigorous evaluation is necessary, and outline core principles for conducting meaningful assessments. This talk covers essential principles, methods (benchmarks, simulation, human feedback), and metrics for evaluating agentic system performance, highlighting key challenges.