Skip to main content

Generalizability of Large Language Model-Based Agents: A Comprehensive Survey

Publication ,  Journal Article
Zhang, M; Yang, Y; Xie, R; Dhingra, B; Zhou, S; Pei, J
Published in: ACM Computing Surveys
July 31, 2026

Large Language Model (LLM)-based agents have recently emerged as a new paradigm that extends the capabilities of LLMs beyond text generation to dynamic interaction with external environments. A critical challenge lies in ensuring their eneralizability – the ability to maintain consistently high performance across varied instructions, tasks, environments, and domains, especially those different from the agent’s fine-tuning data. Despite growing interest, the concept of generalizability in LLM-based agents remains underdefined, and systematic approaches to measure and improve it are lacking. We provide the first comprehensive review of generalizability in LLM-based agents. We begin by clarifying the definition and boundaries of agent generalizability. We then review existing benchmarks. Next, we categorize strategies for improving generalizability into three groups: methods targeting the backbone LLM, targeting agent components, and targeting their interactions. Furthermore, we introduce the distinction between eneralizable frameworks and eneralizable agents and outline how generalizable frameworks can be translated into agent-level generalizability. Finally, we identify future directions, including the development of standardized evaluation frameworks, variance- and cost-based metrics, and hybrid approaches that integrate methodological innovations with agent architecture-level designs. We aim to establish a foundation for principled research on building LLM-based agents that generalize reliably across diverse real-world applications.

Duke Scholars

Published In

ACM Computing Surveys

DOI

EISSN

1557-7341

ISSN

0360-0300

Publication Date

July 31, 2026

Volume

58

Issue

10

Start / End Page

1 / 44

Publisher

Association for Computing Machinery (ACM)

Related Subject Headings

  • Information Systems
  • 46 Information and computing sciences
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Zhang, M., Yang, Y., Xie, R., Dhingra, B., Zhou, S., & Pei, J. (2026). Generalizability of Large Language Model-Based Agents: A Comprehensive Survey. ACM Computing Surveys, 58(10), 1–44. https://doi.org/10.1145/3794858
Zhang, Minxing, Yi Yang, Roy Xie, Bhuwan Dhingra, Shuyan Zhou, and Jian Pei. “Generalizability of Large Language Model-Based Agents: A Comprehensive Survey.” ACM Computing Surveys 58, no. 10 (July 31, 2026): 1–44. https://doi.org/10.1145/3794858.
Zhang M, Yang Y, Xie R, Dhingra B, Zhou S, Pei J. Generalizability of Large Language Model-Based Agents: A Comprehensive Survey. ACM Computing Surveys. 2026 Jul 31;58(10):1–44.
Zhang, Minxing, et al. “Generalizability of Large Language Model-Based Agents: A Comprehensive Survey.” ACM Computing Surveys, vol. 58, no. 10, Association for Computing Machinery (ACM), July 2026, pp. 1–44. Crossref, doi:10.1145/3794858.
Zhang M, Yang Y, Xie R, Dhingra B, Zhou S, Pei J. Generalizability of Large Language Model-Based Agents: A Comprehensive Survey. ACM Computing Surveys. Association for Computing Machinery (ACM); 2026 Jul 31;58(10):1–44.

Published In

ACM Computing Surveys

DOI

EISSN

1557-7341

ISSN

0360-0300

Publication Date

July 31, 2026

Volume

58

Issue

10

Start / End Page

1 / 44

Publisher

Association for Computing Machinery (ACM)

Related Subject Headings

  • Information Systems
  • 46 Information and computing sciences