GraphOmni is the first fully extensible benchmark for evaluating the reasoning capabilities of LLMs on graph-theoretic tasks in natural language. Spanning 6 graph tasks, 7 graph types, 7 serialization formats, and 9 prompt schemes, it systematically explores how these interdependent dimensions shape model performance. Our extensive evaluations reveal that even top models like Claude-3.5 and o4-mini excel on some combinations yet leave substantial room for improvement on harder tasks. To streamline tuning, we introduce a reinforcement learning-inspired framework that adaptively selects optimal configurations at minimal cost. This open, modular platform lays a robust foundation for advancing LLMs’ graph understanding and reasoning.