Agentic AI: A Progression of Language Model Usage

Agentic AI: A Progression of Language Model Usage

On 5 February 2025, Stanford University’s Stanford Online hosted an in‑depth session on Agent AI, revealing how today’s large language models (LLMs) evolve into autonomous agents capable of multi‑step reasoning, tool integration, memory management, and orchestrated sub‑agent collaboration. The talk—delivered by AI researcher Insop Song—mapped the journey from next‑word prediction to sophisticated plan–act–observe loops, highlighting design patterns, real‑world examples, and best practices for building robust AI agents.

Overview of Language Models

Insop Song opened by tracing the foundation of modern LLMs: during pre-training, models consume vast public corpora—web text, books, articles—and learn to predict the most probable next token in any given sequence. Following pre-training, a post-training phase combines instruction-following fine-tuning with reinforcement learning from human feedback, aligning raw predictive capacity with user-oriented goals and improving reliability in conversational settings.

Common Limitations and Mitigation

Despite remarkable fluency, base LLMs suffer from several persistent challenges. They can hallucinate—confidently asserting false or unverifiable facts—and their knowledge remains frozen at the time their training data was collected, leaving them blind to recent events or proprietary information. They likewise offer no built-in attribution for the sources of their outputs, and expanding context windows for richer prompts incurs latency and cost. To address these issues, Insop Song detailed retrieval-augmented generation: by chunking and embedding a target knowledge base (be it proprietary documents or live web data) into a vector store, agents can retrieve only the most relevant passages at query time and graft them into prompts, ensuring that responses draw strictly on verifiable sources.

From Language Models to Autonomous Agents

Transitioning beyond single-turn Q&A, autonomous agents interleave reasoning with action in a continuous plan–act–observe cycle. At each step, the agent first decomposes a high-level goal into subtasks, then invokes external tools or APIs (for weather lookups, database queries, code execution, etc.), ingests the returned observations, and updates its internal memory or conversational state. This cyclical interplay empowers LLMs not only to generate text but also to interact meaningfully with external systems and environments.

Core Design Patterns for AI Agents

Building such agents relies on four intertwined patterns. First, explicit planning prompts the model to break complex tasks into clear, ordered steps. Second, reflection encourages the agent to critique and refine its own outputs before proceeding. Third, tool usage integrates function calling or code execution—allowing real-time data retrieval and precise computations outside the model’s static knowledge. Finally, multi-agent collaboration partitions a problem among specialized sub-agents (each with its own persona or prompt), whose individual outputs are then orchestrated into a coherent whole.

Real-World Applications

These patterns unlock powerful applications across domains. In software development, agents can generate, test, and debug code within sandboxed environments, iteratively refining patches. Research assistants can perform structured web or database searches, synthesize findings, and draft literature reviews or executive summaries. In customer support and task automation, agents maintain memory of past interactions, consult policy databases, and execute order or refund workflows through integrated API calls.

Looking Ahead: The Agent AI Frontier

Agent‑based LLMs mark the next step in AI’s evolution—transitioning from passive text generators into proactive problem solvers. By mastering planning, tool integration, reflection, and multi‑agent patterns, developers can unlock complex automations across industries. As compute‑at‑inference (“test‑time scaling”) and reasoning‑optimized models mature, the agent paradigm will only grow more capable—empowering non‑expert creators to build AI workflows as effortlessly as TikTok democratized video editing.

A recording, slides, and code references will be shared soon via Stanford Online. For further resources and upcoming AI seminars, visit Stanford’s Center for Global Education Generative AI catalog.


Stanford Online is Stanford University’s digital learning hub, offering on-demand courses, professional certificates, and full degree programs across disciplines—from data science and leadership to design thinking—through self-paced and instructor-led formats that bring Stanford’s research-driven curriculum and faculty expertise to learners worldwide.

The Conf is a platform that reports on scholarly conferences, symposia, roundtables, book talks, and other academic events. It is managed by a group of students from leading American and European universities and is published by Alma Mater Europaea University, Location Vienna.

Share this article: