Documente Academic
Documente Profesional
Documente Cultură
2006/3/21
2.1 Define in your own words the following terms: agent, agent function, agent
program, rationality, autonomy, reflex agent, model-based agent, goal-based agent,
utility-based agent, learning agent.
2.5 For each of the following agents, develop a PEAS description of the task
environment:
a. Robot soccer player; b. Internet book-shopping agent; c. Autonomous Mars rover;
d. Mathematician’s theorem-proving assistant.
1
Autonomous Mars Collect, Mars, vehicle Collection , Video camera,
rover analyze and analysis, and audio
explore motion receivers,
samples on devices, radio communication
Mars transmitter links
Theorem-proving Time The theorem Accept the Input device
assistant requirement, to prove, right theorem, that reads the
degree of existing reject the theorem to
correction axioms wrong prove
theorem, infer
based on
axioms and
facts
c. The above randomized agent will perform poorly in environments like the
following one. It will need a lot of time to get through the long passage because of the
random walk.
2
d. A reflex agent with state can first explore the environment thoroughly and build a
map of this environment. This agent can do much better that the simple reflex agent
because it maintains the map of the environment and can choose action based on not
only the current percept, but also current location inside the map.
2.12 The vacuum environments in the preceding exercises have all been deterministic.
Discuss possible agent programs for each of the following stochastic versions:
a. Murphy’s law: twenty-five percent of the time, the Suck action fails to clean the
floor if it is dirty and deposits dirt onto the floor if the floor is clean. How is your
agent program affected if the dirt sensor gives the wrong answer 10% of the time?
b. Small children: At each time step, each clean square has a 10% chance of becoming
dirty. Can you come up with a rational agent design for this case?
a. The failure of Suck action doesn’t cause any problem at all as long as we replace
the reflex agent’s ‘Suck’ action by ‘Suck until clean’.
If the dirt sensor gives wrong answers from time to time, the agent might just leave
the dirty location and maybe will clean this location when it tours back to this location
again, or might stay on the location for several steps to get a more reliable
measurement before leaving. Both strategies have their own advantages and
disadvantages. The first one might leave a dirty location and never return. The latter
might wait too long in each location.
b. In this case, the agent must keep touring and cleaning the environment forever
because a cleaned square might become dirty in near future.