Updated:

Intelligent Agents

  • Chess: logical, well-organized problem with finite life time and objective.
  • Medical diagnosis: ill-defined problem, where environment is changing. (multiple diseases, patient may not reveal everything, etc.)
  • Walking on two legs: real-time control, getting through obstacles, maintaining stability, etc.
  • Taxi driver: complex, quickly evolving environment.

How do we characterize These different types of intelligent agents?

4 properties of the Agent model

(Characterise requirements for an agent in terms of its percepts, actions, environment, and performance measure.)

  • **Percepts/observations** of the environment
    • Made by sensors (How does the model sense the environment?)
    • Input
    • How does the model sense the enviroment?
  • **Actions** which may affect the environment
    • Made by actuators (What action does the agent perform?)
    • Output
    • i.e. Ability to exelerate, move, accept money, change the state of a chess board, etc.
  • **Environment** in which the agent exists
    • What’s being sensed and modified?
    • More complex, need to be modeled.
  • **Performance measure** of the desirability of environment states
    • What’s the object applied to the agent?

  • Example: Automated taxi
    • Percepts: video, accelerometers, temperature gauges, engine sensors, keyboard, GPS, …
    • Actions: steer, accelerate, brake, horn, speak/display, …
    • Environment: city streets, freeways, traffic (other vehicles, traffic lights), pedestrians, weather, customer, …
    • Performance measure: safety, reach destination, maximize profits, obey laws, passenger comfort, …

  • Agents can be evaluated empirically, sometimes analysed mathematically.
  • Agent is a function from percept sequences to actions.
  • Ideal rational agent picks actions that is expected to maximise its performance measure (based on the percept sequence and its built-in knowledge)
    • Rational != omniscent (know everything)
    • Rational != clairvoyant (automatically predict the future state)
    • Rational != successful (need to experiment, then do the best it can)

4 Agent types

What is the functionality that lies inside the agent? (Choose and justify choice of agent type for a given problem.)

  • **Simple reflex agents**

    ai l2

    • Assumes that the world is fully observable (no missing information)
    • No long term reasoning (may result to infinite loops)
    • Tasks where this type can be sufficient: automatic door (detects person in front of the sensor)
  • **Model-based reflex agents**

    ai l2-1

    • Assumes some of the world is not observable, and uses a model to infer the state of the environment
    • Long-term decision, but still needs improvement
  • **Goal-based agents**

    image

    • Looks ahead to achieve desirable future state
    • Searches for the best action to get to the focused, specific, single goal
  • **Utility-based agents**

    image

    • Dealing with different goals (short & long)
    • Utility: contains a measure of how desirable a future state is
      • Consider which one would maximize my happiness
      • Mapping state to a value (optimize the expected outcome)
      • More general, thinks about the trade-off between achieving different goals
    • Multiple goals when trying to maximize the profit in a taxi
      • Keeping the customer happy
      • Getting the customer to the destination
      • Obeying the traffic rules (avoiding fines)
      • Not taking a route that is unnecessarily long

5 Environment types

You should choose the agent type depending on the environment. (characterise the environment for a given problem.)

Environments may or may not be:

  • **Observable**
    • Percept contains all relevant information about the world; Everything is in view.
    • partially observable: Environments cannot be directly sensed via percept.
  • **Deterministic**
    • Current state of the world uniquely determines the next.
    • Having high certainty about the consequences (i.e. I will get to the beach by 3:30pm if I take this train.)
    • stochastic: The next state is randomly determined.
  • **Episodic**
    • Only the current (or recent) percept is relevant. (Only consider short-term actions)
    • sequential: The consequences are long-term.
  • **Static**
    • Environment doesn’t change while the agent is deliberating.
    • (i.e. chess: board does not change while the agent makes a move.)
    • dynamic: environment changes while making the decision.
      • (i.e. driving cars: environment changes constantly.)
  • **Discrete**
    • There are finite number of possible percepts/actions.
    • (i.e. chess: number of squares on the board, number/state of pieces are finite.)
    • continuous: infinite number of possible percepts/actions.
      • (i.e. driving a car)

image

  • The environment type largely determines the agent design.
  • The real world is partially observable, stochastic, sequential, dynamic, and continuous.

Leave a comment