|
563 | 563 | "A model-based reflex agent maintains some sort of **internal state** that depends on the percept history and thereby reflects at least some of the unobserved aspects of the current state. In addition to this, it also requires a **model** of the world, that is, knowledge about \"how the world works\".\n",
|
564 | 564 | "\n",
|
565 | 565 | "The schematic diagram shown in **Figure 2.11** of the book will make this more clear:\n",
|
566 |
| - "<img src=\"files/images/model_based_reflex_agent.jpg\">" |
| 566 | + "<img src=\"images/model_based_reflex_agent.jpg\">" |
567 | 567 | ]
|
568 | 568 | },
|
569 | 569 | {
|
|
650 | 650 | "A goal-based agent needs some sort of **goal** information that describes situations that are desirable, apart from the current state description.\n",
|
651 | 651 | "\n",
|
652 | 652 | "**Figure 2.13** of the book shows a model-based, goal-based agent:\n",
|
653 |
| - "<img src=\"files/images/model_goal_based_agent.jpg\">\n", |
| 653 | + "<img src=\"images/model_goal_based_agent.jpg\">\n", |
654 | 654 | "\n",
|
655 | 655 | "**Search** (Chapters 3 to 5) and **Planning** (Chapters 10 to 11) are the subfields of AI devoted to finding action sequences that achieve the agent's goals.\n",
|
656 | 656 | "\n",
|
|
659 | 659 | "A utility-based agent maximizes its **utility** using the agent's **utility function**, which is essentially an internalization of the agent's performance measure.\n",
|
660 | 660 | "\n",
|
661 | 661 | "**Figure 2.14** of the book shows a model-based, utility-based agent:\n",
|
662 |
| - "<img src=\"files/images/model_utility_based_agent.jpg\">" |
| 662 | + "<img src=\"images/model_utility_based_agent.jpg\">" |
663 | 663 | ]
|
664 | 664 | },
|
665 | 665 | {
|
|
673 | 673 | "A learning agent can be divided into four conceptual components. The **learning element** is responsible for making improvements. It uses the feedback from the **critic** on how the agent is doing and determines how the performance element should be modified to do better in the future. The **performance element** is responsible for selecting external actions for the agent: it takes in percepts and decides on actions. The critic tells the learning element how well the agent is doing with respect to a fixed performance standard. It is necesaary because the percepts themselves provide no indication of the agent's success. The last component of the learning agent is the **problem generator**. It is responsible for suggesting actions that will lead to new and informative experiences. \n",
|
674 | 674 | "\n",
|
675 | 675 | "**Figure 2.15** of the book sums up the components and their working: \n",
|
676 |
| - "<img src=\"files/images/general_learning_agent.jpg\">" |
| 676 | + "<img src=\"images/general_learning_agent.jpg\">" |
677 | 677 | ]
|
678 | 678 | }
|
679 | 679 | ],
|
|
0 commit comments