Recap of Lecture 2

The Four Pillars of Rationality

Rationality is defined by four core components that determine an agent's success:

Environment Properties

Environments are categorized by several key characteristics that influence agent design:


Agents Vs Objects


Agent Architectures

The lecture details four basic agent types, ordered by increasing complexity and generality:

1. Simple Reflex Agents

A simple reflex agent can fall into an infinite loop when it operates in a partially observable environment.

A specific case presented in the lecture involves a vacuum cleaner agent that does not observe its current location:

A possible solution to break such infinite loops in simple reflex agents is to randomize the action.


2. Model-based Reflex Agents


The main difference between a Simple Reflex Agent and a Model-based Reflex Agent lies in how they represent and interact with their environment:


3. Goal-based Agents


4. Utility-based Agents

Learning Agents

All the above architectures can be transformed into learning agents. A learning agent consists of four main components:


The following table compares the different agent architectures presented in the lecture, ordered by their increasing level of generality and complexity.

Agent Type Core Principle Key Characteristics & Pros Cons & Limitations
Simple Reflex Maps current percepts directly to actions using condition-action rules. Simple to design; fast reaction times; memoryless. Limited intelligence; ignores percept history; fails in partially observable environments; prone to infinite loops, cannot learn or adapt to new situations.
Model-based Reflex Maintains an internal state (model) to track parts of the world not currently visible. Uses transition and sensor models to predict how the world evolves and how actions affect it. More complex than simple reflex; requires constant updating of the internal world state.
Goal-based Acts to achieve specific, explicit goals using search and planning. Goal-oriented; flexible behavior; can solve complex problems by predicting consequences. Planning is computationally expensive; struggles with incomplete information; cannot adapt to new situations without learning.
Utility-based Uses a utility function to measure the "happiness" or "success" of a state. Makes rational decisions by maximizing expected utility; handles trade-offs (e.g., speed vs. safety) and uncertainty. Challenging to define accurate utility functions; high computational cost for calculating expected utility.
Learning Can be applied to any architecture to improve performance over time. Includes a learning element, critic (evaluator), and problem generator for exploration. Exploration can be costly in the short term (e.g., fewer tips for a taxi driver while experimenting).

Another Perspective