TU Wien:Einführung in die Künstliche Intelligenz VU (Eiter, Tompits)/Prüfung 2012-10-02 Ausarbeitung
Fragen übernommen aus dem Informatik-Forum.
Planning[Bearbeiten | Quelltext bearbeiten]
Write a STRIPS action[Bearbeiten | Quelltext bearbeiten]
Drive(car, from, to) PRECOND (at(car, from), car(car), Location(from), location(to)) EFFECT (at(car, to), not at(car, from))
What is a consistent plan? What is a solution?[Bearbeiten | Quelltext bearbeiten]
A consistent plan is: A plan with no cycles in the ordering constraints and no conflicts with the causal links
A solution is: A consistent plan with no open preconditions
Describe the two possible ways in state space search[Bearbeiten | Quelltext bearbeiten]
- progression planning (forward state-space search) - from initial state to goal
- regression planning (backward state-space search) - from goal to initial state
Multiple choice ADL/STRIPS[Bearbeiten | Quelltext bearbeiten]
- Does ADL allow equality? - Yes
- Does STRIPS allow Conjunctions in goals? - Yes, but no disjunction.
Given a Lottery L, ...[Bearbeiten | Quelltext bearbeiten]
A expected monetary value EMV(L) and Utility Function U(L), a risk-averse agent would
- prefere U(L) > U(S(EMV(L))
- prefere U(L) < U(S(EMV(L))
where S(EMV(L)) is the state of having money EMV(L).
Describe 3 Axioms of Utility theory: Orderability, Continuity and Monotonicity[Bearbeiten | Quelltext bearbeiten]
- Orderability: Given any two lotteries, a rational agent cannot avoid deciding which one it prefers, or whether it is indifferent between them.
- Exactly one of holds.
- Continuity: If some lottery B is between A and C in preference, then:
- there is some probability p for which the agent will be indifferent between getting B for sure and the lottery that yields with probability and with probability .
- there is some probability p for which the agent will be indifferent between getting B for sure and the lottery that yields with probability and with probability .
- Monotonicity: Suppose two lotteries have the same possible outcomes and .
- If an agent prefers to , then the agent must prefer precisely the lottery that has a higher probability for outcome .
Describe the components of a decision network[Bearbeiten | Quelltext bearbeiten]
- Chance nodes (ovals): represent random variables.
- E.g., the agent is uncertain about construction costs, the level of air traffic, the potential for litigation.
- There are also the Deaths, Noise, and Cost variables, depending on the site chosen.
- Chance nodes are associated with a conditional probability distribution that is indexed by the state of parent nodes.
- Decision nodes (rectangles): represent points where decision maker has a choice of actions; e.g., the choice of an airport site influences the cost, noise, etc.
- Utility nodes (diamonds): represent the agent’s utility function.
- It has as parents all variables describing the outcome that directly affect utility.
Searches[Bearbeiten | Quelltext bearbeiten]
a small tree is given. Give the order in that nodes are expanded for all 5 uninformed search algorithms (5 points)[Bearbeiten | Quelltext bearbeiten]
Given admissible heuristics (h1, h2,... hn). Which possible (new) heuristic is best?[Bearbeiten | Quelltext bearbeiten]
Admissible heuristics have a problem with graph search. Which? How can this be changed?[Bearbeiten | Quelltext bearbeiten]
Even if h is admissible, A*-Search can reach goals in graph search with non optimal costs. To fix this problem we can change the algorithm and add complicated bookkeeping (might ruin run time) or impose a stronger restriction on h: consistence (f-value is non-decreasing on every path)
Learning from Observation[Bearbeiten | Quelltext bearbeiten]
What is inductive learning?[Bearbeiten | Quelltext bearbeiten]
Given a training set (x1, y1) ... (xn, yn) of examples find a function h that approximates f. h is then called hypothesis and must be from a restricted class of functions (hypothesis space). If h = f is possible then the learning problem is realizable. Consider Ockham's razor when creating this function "maximise simplicity under consistency"
What is a learning curve? Which correlation is described by it? Which two circumstances have negative effect on the learning curve?[Bearbeiten | Quelltext bearbeiten]
% correct on test set as a function of training set size (x-axis: size of training set, y-axis: % correct on test set) Describes correlation between test set size and % of correct outcomes Depends on realizability (non realizable through missing attributes or too restrictive hypothesis space) and redundant expressiveness (due to loads of irrelevant attributes)
Describe a learning agent + diagram[Bearbeiten | Quelltext bearbeiten]
learning element: make improvements
performance element: select external actions
critic: performance, result assessment
problem generator: suggest actions for new experiences
Construct a neural network that represents a half adder: two inputs (I1, I2), two outputs (C = I1 & I2, S = I1 xor I2)[Bearbeiten | Quelltext bearbeiten]
Advantages & Disadvantages of neural networks (or learning by observation)[Bearbeiten | Quelltext bearbeiten]
PROs
- less need for determining relevant input factors
- inherent parallelism
- easier to develop than statistical methods
- capability to learn and improve
- good for complex pattern recognition tasks
- usable for unstructured and difficult input (like images)
- fault tolerance
CONs
- choice of parameters (layers, units) requires skill
- requires sufficient training material
- resulting hypothesis cannot be understood easily
- knowledge implicit (subsymbolic representation)
- verification, behaviour prediction difficult