Artificial Intelligence/Intelligent Agents: Difference between revisions
Brodriguez (talk | contribs) (Create page) |
Brodriguez (talk | contribs) (Update page) |
||
Line 7: | Line 7: | ||
* Sensor - A device that takes some form of input from the surrounding environment. | * Sensor - A device that takes some form of input from the surrounding environment. | ||
* Actuator - A device that achieves physical movement of some kind, by converting energy into mechanical force. | * Actuator - A device that achieves physical movement of some kind, by converting energy into mechanical force. | ||
* Percept - An agent's perceptual inputs at a given time. | |||
* Percept Sequence - The complete history of everything the agent has perceived, in order. | |||
* Agent Function - A mathematical function that maps behavior to percept sequences. | |||
** This "Agent Function" is effectively the theoretical implementation of the agent. | |||
* Agent Program - A concrete implementation of an agent, within some physical system. | |||
** This "Agent Program" is effectively the practical implementation of the agent. | |||
Line 40: | Line 46: | ||
=== Types of Actuators === | === Types of Actuators === | ||
An actuator is anything that exerts some sort of force on the agent or the surrounding environment. | |||
For example, an agent might have an actuator in the form of: | |||
* A robotic arm that grabs or manipulates objects. | |||
* A wheel for a self-driving car, which allows the car to move. | |||
== Evaluating Performance == | == Evaluating Performance == | ||
We generally use objective, quantifiable metrics to describe the level of success for a given agent's behavior.<br> | |||
The exact details of this measurement is often called "Performance Measures". | |||
For AI meant to have [[Programming/Artificial Intelligence#Rational AI Systems | rational behavior]], we tend to measure based on the following: | |||
* Prior knowledge of environment. | |||
* Possible actions that can be performed. | |||
* Precept sequence at the time of the action that was taken. | |||
Ideally, for each possible percept sequence, the agent will take an action that will maximize objective success.<br> | |||
Prior knowledge, as well as current understanding of the environment, can play a large role in this. | |||
=== Agent Knowledge Properties === | |||
Below are various properties that affect an agent's sensory input and overall environmental knowledge. | |||
Environments can be "Fully Observable", "Partially Observable", or "Unobservable": | |||
* Fully Observable - Sensors can detect all aspects of environment that are relevant to the agent's actions. | |||
* Unobservable - Sensors are broken or otherwise unable to gather data relevant to the agent's actions. | |||
* Partially Observable - Some combination of the above two states. | |||
Sensor data can be "Episodic" or "Sequential": | |||
* Episodic - One precept leads to exactly one action. | |||
** Agents with episodic sensory data tend to not retain memory of past precepts. | |||
** For example, maybe a robotic arm at a factory. It only cares if an item is present or not at the given moment. | |||
* Sequential - Multiple precepts (aka, multiple moments in time) are used together to determine an action. | |||
** Agents with sequential sensory data have to retain some memory of past precepts. | |||
** For example, a self-driving car needs to, at minimum, retain recent history of the past few minutes, in order to determine road state, nearby traffic, etc. | |||
On taking an action, the resulting state can have one of three properties: | |||
* Deterministic - The next state is always determined with 100% certainty, by a combination of the current state plus the agent's action. | |||
** In other words, upon taking an action, the next state is always fully predictable. | |||
** Ex: A robotic arm in a controlled factory setting. It will know what the resulting state of the environment will be, before taking an action. | |||
* Non-deterministic - The next state is not determinable by the agent's action. The agent will need to take an action and then check the environment to find out the current state. | |||
** Ex: A self-driving car may turn left. But that only tells the car what it's doing. It has no idea what other entities on the road will do, regardless of its action. | |||
* Stochastic - The next state is uncertain, but quantifiable by probabilities. | |||
** Ex: If a vacuum cleaning a square has a 90% chance to clean, and a 10% chance to spread debris to nearby locations. | |||
*** In such a case, the next state is uncertain, but the agent can make intelligent assumptions, based on the likely-hood of what will happen from each action. | |||
Environment can be one of several states, as an agent determines its next action: | |||
* Static - The environment does not change while the agent considers its next action. | |||
** Ex: An agent designed to play a game of solitaire. The board will not change until the agent takes an action. | |||
* Dynamic - The environment may change while the agent considers its next action. | |||
** Ex: A self-driving car. Nearby entities will act, regardless of how long the current agent takes to determine an action. | |||
* Semi-Dynamic - The environment technically does not change while the agent considers its next action. BUT the agent's overall performance score can change depending on how long it takes to decide. | |||
Related to above: | |||
* Discrete Time - The environment is split into disparate steps. | |||
** Ex: A card game. | |||
* Continuous Time - The environment is a continuous string of input. | |||
** Ex: A self-driving car on the road. | |||
Lastly, the number of other agents in the environment also make an impact: | |||
* Single Agent Environment - Only one agent exists in the environment. No consideration is needed for other entities. | |||
* Multi-Agent Environment - Two or more agents exist in the same environment. They may or may not need to account for each other in their actions. | |||
** Multi-Agent Environments can be further divided into the following: | |||
*** Cooperative Multi-Agent Environment - Two or more agents exist in the same environment, and work together to achieve some goal. | |||
*** Competitive Multi-Agent Environment - Two or more agents exist in the same environment, and directly oppose each other in achieving goals. |
Revision as of 18:23, 5 October 2021
This page discusses the general ideas, and overall implementation of an intelligent agent.
Terminology
- PEAS - An abbreviation to remember the "basics" of a functioning agent.
- Stands for "Performance Measures, Environment, Actuators, Sensors"
- Sensor - A device that takes some form of input from the surrounding environment.
- Actuator - A device that achieves physical movement of some kind, by converting energy into mechanical force.
- Percept - An agent's perceptual inputs at a given time.
- Percept Sequence - The complete history of everything the agent has perceived, in order.
- Agent Function - A mathematical function that maps behavior to percept sequences.
- This "Agent Function" is effectively the theoretical implementation of the agent.
- Agent Program - A concrete implementation of an agent, within some physical system.
- This "Agent Program" is effectively the practical implementation of the agent.
What is an Agent?
As mentioned in Artificial Intelligence, an "Intelligent Agent" is an entity that can act.
It generally acts to achieve the "best possible" outcome, and uses information from the surrounding environment in some manner.
To elaborate, an Intelligent Agent acquires environmental information via some form of sensors.
And it often acts upon that environment through some sort of actuators.
Said actions will depend on the agent's programming, as well as the current input from its sensors.
It may process this input using one (or multiple) of the following:
- Reasoning via probabilities and statistics.
- Decisions via Bayesian Networks, analysis of observation (sensory) sequences, or other methods.
- Learning via training from statistical models, neural networks, and other data-driven methods.
Types of Sensors
Sensors are anything which detect or otherwise mimic input of the 5 human senses (sight, touch, smell, hearing, taste), or even other values that humans can't inherently detect. Includes:
- Cameras - Mimics "sight" in the form of continuous image sequences.
- Microphones - Mimics "sound" in the form of wavelength sequences.
- Thermometers - Mimics "touch" in the form of detecting temperature.
- Current & Voltage Sensors - Detects electricity through a given circuit.
- Chemical & Gas Sensors - Detects environmental composition.
- Etc
What sensor an AI uses heavily depends on the AI's application, as well as the environment it will reside in.
For example, an AI that exists within a closed, indoor space may never need a thermometer.
Meanwhile, a self-driving car AI may need a thermometer to help detect the type of weather its driving in.
Types of Actuators
An actuator is anything that exerts some sort of force on the agent or the surrounding environment.
For example, an agent might have an actuator in the form of:
- A robotic arm that grabs or manipulates objects.
- A wheel for a self-driving car, which allows the car to move.
Evaluating Performance
We generally use objective, quantifiable metrics to describe the level of success for a given agent's behavior.
The exact details of this measurement is often called "Performance Measures".
For AI meant to have rational behavior, we tend to measure based on the following:
- Prior knowledge of environment.
- Possible actions that can be performed.
- Precept sequence at the time of the action that was taken.
Ideally, for each possible percept sequence, the agent will take an action that will maximize objective success.
Prior knowledge, as well as current understanding of the environment, can play a large role in this.
Agent Knowledge Properties
Below are various properties that affect an agent's sensory input and overall environmental knowledge.
Environments can be "Fully Observable", "Partially Observable", or "Unobservable":
- Fully Observable - Sensors can detect all aspects of environment that are relevant to the agent's actions.
- Unobservable - Sensors are broken or otherwise unable to gather data relevant to the agent's actions.
- Partially Observable - Some combination of the above two states.
Sensor data can be "Episodic" or "Sequential":
- Episodic - One precept leads to exactly one action.
- Agents with episodic sensory data tend to not retain memory of past precepts.
- For example, maybe a robotic arm at a factory. It only cares if an item is present or not at the given moment.
- Sequential - Multiple precepts (aka, multiple moments in time) are used together to determine an action.
- Agents with sequential sensory data have to retain some memory of past precepts.
- For example, a self-driving car needs to, at minimum, retain recent history of the past few minutes, in order to determine road state, nearby traffic, etc.
On taking an action, the resulting state can have one of three properties:
- Deterministic - The next state is always determined with 100% certainty, by a combination of the current state plus the agent's action.
- In other words, upon taking an action, the next state is always fully predictable.
- Ex: A robotic arm in a controlled factory setting. It will know what the resulting state of the environment will be, before taking an action.
- Non-deterministic - The next state is not determinable by the agent's action. The agent will need to take an action and then check the environment to find out the current state.
- Ex: A self-driving car may turn left. But that only tells the car what it's doing. It has no idea what other entities on the road will do, regardless of its action.
- Stochastic - The next state is uncertain, but quantifiable by probabilities.
- Ex: If a vacuum cleaning a square has a 90% chance to clean, and a 10% chance to spread debris to nearby locations.
- In such a case, the next state is uncertain, but the agent can make intelligent assumptions, based on the likely-hood of what will happen from each action.
- Ex: If a vacuum cleaning a square has a 90% chance to clean, and a 10% chance to spread debris to nearby locations.
Environment can be one of several states, as an agent determines its next action:
- Static - The environment does not change while the agent considers its next action.
- Ex: An agent designed to play a game of solitaire. The board will not change until the agent takes an action.
- Dynamic - The environment may change while the agent considers its next action.
- Ex: A self-driving car. Nearby entities will act, regardless of how long the current agent takes to determine an action.
- Semi-Dynamic - The environment technically does not change while the agent considers its next action. BUT the agent's overall performance score can change depending on how long it takes to decide.
Related to above:
- Discrete Time - The environment is split into disparate steps.
- Ex: A card game.
- Continuous Time - The environment is a continuous string of input.
- Ex: A self-driving car on the road.
Lastly, the number of other agents in the environment also make an impact:
- Single Agent Environment - Only one agent exists in the environment. No consideration is needed for other entities.
- Multi-Agent Environment - Two or more agents exist in the same environment. They may or may not need to account for each other in their actions.
- Multi-Agent Environments can be further divided into the following:
- Cooperative Multi-Agent Environment - Two or more agents exist in the same environment, and work together to achieve some goal.
- Competitive Multi-Agent Environment - Two or more agents exist in the same environment, and directly oppose each other in achieving goals.
- Multi-Agent Environments can be further divided into the following: