This repo contains code snippets used for my 2025 lecture at Columbia University, an introduction to AI Agents (slides here) titled: "From LLMs to agents: The road to autonomy is paved with good intentions".
We use uv to manage the environment. To set up the environment, run:
uv syncWe use Anthropic APIs for the LLM backend. Make sure to create a .env file in src directory using local.env as a template. If you prefer to use other models, we rely on LiteLLM interface, so it should be straightforward to swap out the model backend. Make sure to change both the .env and the LLM init accordingly.
To run the code snippets, navigate to the src directory and execute the desired Python script with uv, for example:
cd src
uv run text_to_tool_to_text.pyPlease refer to the slides for the general context and the AI background. The code has been developed for educational purposes only, using Claude Code as an AI assistant throughout the development process.
A standalone Fast MCP server can be found in the fast_mcp directory, to showcase how to build and leverage a minimal server inside of an agentic loop: the server gets automatically called from the agent_with_mcp.py script, but could also be run in a standalone fashion.
This project is provided "as is" and it is licensed under the MIT License - see the LICENSE file for details.