Analogical Learning and Problem Solving
1 Aug 2011
Ramakrishna, 114054001Abhijit, 114056002Anup, 10305915
Analogy
• The analogous solution to the radiation problem is to simultaneously direct multiple low-intensity rays toward the tumor from different directions. In this way the healthy tissue will be left unharmed, but the effects of the multiple low-intensity rays will summate and destroy the tumor
What is it ?
• Machine Learning– Supervised Learning
• Inductive Learning– Decision Trees, Version Space etc.
• Analogical Learning
– Unsupervised Learning– Reinforcement Learning
• Intuition– If two entities are similar in some respects then they
could be similar in other respects as well
Steps in Analogical Problem Solving
• Identifying Analogy– Find a known/experienced problem that is analogous to the input
problem
• Mapping– Relevant parts of the experienced problem are selected and
mapped on to the input problem to transform the new problem and derive some hypothesis for the solution
• Validation– The correctness of the newly constructed solution is validated
using theorem or simulation etc.
• Learning– If the validation is found to work well, the new knowledge is
encoded and saved for future usage.
Analogy Representation
• Appropriate level of “analogical structures” needs to be represented– Too Detailed or Too Abstract representation
will make it had to draw analogy
• Example
Analogical Structures1. A fortress was located in the center of the
country.2. Many roads radiated out from the fortress.3. A general wanted to capture the fortress
with his army.4. The general wanted to prevent mines on
the roads from destroying his army and neighboring villages.
5. As a result the entire army could not attack the fortress along one road.
6. However, the entire army was needed to capture the fortress.
7. So an attack by one small group would not succeed.
8. The general therefore divided his army into several small groups.
9. He positioned the small groups at the heads of different roads.
10. The small groups simultaneously converged on the fortress.
11. In this way the army captured the fortress.
1. A tumor was located in the interior of a patient's body.
2. -3. A doctor wanted to destroy the tumor with
rays. 4. The doctor wanted to prevent the rays from
destroying healthy tissue. 5. As a result the high-intensity rays could not
be applied to the tumor along one path.6. However, high-intensity rays were needed
to destroy the tumor. 7. So applying one low-intensity ray would not
succeed. 8. The doctor therefore divided the rays into
several low intensity rays. 9. He positioned the low intensity rays at
multiple locations around the patient's body.
10. The low-intensity rays simultaneously converged on the tumor.
11. In this way the rays destroyed the tumor.
Predicate Analogy• Sample Predicates
– locate(fortress, centre(country))– locate(tumour, interior(body))
– desire(general, capture(army, fortress))– desire(doctor, destroy(rays, tumour))
• Mapping– fortress → tumour– general → doctor– army → rays
• Inference– divide(general, small_group(army))– divide(doctor, small_group(rays))
– position(small_group(army), different_direction)– position(small_group(rays), different_direction)
– simultaneous(converge(small_group(army), fortress))– simultaneous(converge(small_group(rays),tumour))
– destroy(army, fortress)– destroy(rays, tumour)
Analogical Problem Solving
• Needs sophisticated techniques for– Representation of the problem and solution– Matching input problem with existing problem– Concluding inferences
• Very hard problem
• Let us see one system - EUREKA
EUREKA• An analogical problem solving architecture
presented by RANDOLPH M. JONES and PAT LANGLEY
• Quick look at EUREKA– Existing problem
– New problem. Solve by analogy to previous problem
• Consists of– Memory Module– Problem Solving Engine
A B C
A
B C
E F G
E
F HH
G
Memory Representation• As Semantic Network• Example network for block movement problem
Ref: Randolph Jones; Problem Solving via Analogical Retrieval and Analogical Search Control
State100
Goal17
Trans1 Transforms to a new state which satisfies the goal
Problem Solving Engine
• Uses Flexible Means-Ends Algorithm
• Two recursive functions are used– TRANSFORM – APPLY
State X State Z
Op1 State P
State C
Transform and Apply ProceduresTRANSFORM(StateX,Conditions): Returns StateZ If StateX satisfies Conditions Then Return StateZ as StateX Else Let Operator be SELEC_OPERAT0R(StateX,Conditions); If Operator is empty Then Return StateZ as "Failed State1' Else Let StateY be APPLY(StateX,Operator); If StateY is "Failed State" Then Return StateZ as "Failed State" Else Return StateZ as TRANSFORM(StateY,Conditions)
APPLY(StateX,Operator): Returns StateY Let P be PRECONDITIONS(Operator); If StateX s a t i s f i e s P Then Return StateY as EXECUTE(StateX,Operator) Else Let StateW be TRANSFORM(StateX,P); If StateW i s "Failed State" Then Return StateY as "Failed State" Else Return StateY as APPLY(StateW,Operator)
Ref: Randolph Jones; Problem Solving via Analogical Retrieval and Analogical Search Control
Example
Ref: Randolph Jones; Problem Solving via Analogical Retrieval and Analogical Search Control
• Transforming from the initial state (State100) to the final state (State200) which satisfies the given goal (Goal17)
Identifying Analogy and Problem Solving
• Collect some analogical TRANSFORM goals from the existing solution– Use spreading techniques from the current state and
goal nodes– Eg: Trans1, Trans10
• Select the best matching goals– Degree of match– History data (success/failure)
• Make concept analogy– E → A; F → B
• Proceed as per the analogical TRANSFORM goal
Identifying Analogy and Problem Solving Contd…
State500 Goal99
On5 On6
E F
On..
Closest Matching TRNSFORMATION
Goal
……….
EUREKA – Initial Stage
• EUREKA started without any knowledge in its long-term memory– For eg. First Time– never be able to solve any problems, because
there would be no previously solved problems on which to base future decisions
• Operators are stored in the form of simple problems
Case Based Reasoning (CBR) Reasoning method that uses specific past
experiences rather than a corpus of general knowledge.
Problem solving by analogy Recognizing its similarity to a specific known
problem Transferring the solution of the known problem to
the new one. Form of intra-domain analogy.
CBR terminologies Case : Denotes a problem situation Past case : A previously experienced situation which
has been captured and learned for reuse. Case based reasoning : A cyclic and integrated
process of solving a new problem.
Learning in CBR Natural by-product of problem solving. The experience of a successfully solved is retained
inorder to solve future problems If an attempt to solve problem fails, reason of failure
is remembered in order to avoid same mistake in future.
CBR Cycle RETRIEVE
Retrieve one or more previously experienced cases. REUSE
Reusing the case in one way or another. REVISE
Revising the solution based on reusing a previous case.
RETAIN Retaining the new experience by incorporating it
into existing knowledge base.
CBR Cycle
Fig . Taken from: Agnar Aamodt, Enric Plaza, “Case-Based Reasoning Foundational Issues, Methodological Variations, and System Approaches”, AI communications, 1994.
How “cases” are represented ?
Figure taken from “Agnar Aamodt and Enric Plaza, “Case-Based Reasoning: Foundational Issues, Methodological Variations, and System Approaches”,AI Communications, Vol. 7 Nr. 1, March 1994, pp 39-59”
How “cases” are represented ? In dynamic memory model, cases that share similar properties are
organized under a more general structure known as generalized episode – GE.
A generalized episode contains
Norms
denote features common to all cases indexed under a GE. Indices
are features that discriminate between a GE's cases may point to a more specific generalized episode, or directly
to a case. An index is composed of
Index name Index value
Cases
How to find a similar problem ?
When a new case description is given and the best matching is searched for, the input case structure is 'pushed down' the network structure, starting at the root node.
When one or more features of the case match one or more features of a GE, the case is further discriminated based on its remaining features.
Eventually, the case with most features in common with the input case is found.
How a new case is trained to system ?
During storing of a new case, it is discriminated by indexing it under different indices below its most specific generalized episode.
If, during the storage of a case - two cases (or two GEs) end up under the same index, a new generalized episode is automatically created.
An example...
Generalized Episode 2
Norms: Fever
Indices:
1 2 3
Influenza symptoms
Swine Flu Symptoms
Malaria symptoms
Values:
Index:
Influenza
Swine Flu
MalariaCases:
Generalized Episode 3
Norms: Dysentry
Indices:
1 2 3
food poisoning symptoms
Cholera Symptoms
Amoebicsymptoms
Values:
Index:
food poisoning
Cholera Amoebic dysentry
Cases:
Generalized Episode 1
Norms: Disease symptoms
Indices:
1 2
Fever symptoms
Dysentry symptoms
Values:
Index:
An example... Consider a doctor patient senario.
The case of diseases trained are already stored in the tree fashion.
It has got Generalized Episodes(GE) as the entity.
Suppose a new patient comes with malaria symptoms.
Current case symptoms are compared with the already existing ones. So it is directed to the left subtree. Here norms give the features common to that GE.
Then cases are distinguished according to the symptoms which are specific to those.
Ultimately the current case is directed to the index 3 where it is matched and the solution to the already stored case is applied to the current case.
Protos Case-based problem solving and learning system for
heuristic classification tasks.
In Protos, a concept ci is represented extensionally as a collection of examples (called exemplars or cases):
ci = {ei1, ei2, ...}.
The classification and learning algorithmInput: a set of exemplar-based categories C = {c1, c2, ... , cn} and a case (NewCase)
to classify.
REPEAT
Classify:
Find an exemplar of ci є C that strongly matches NewCase and classify NewCase as ci.
Explain the classification.
Learn:
If the expert disagrees with the classification or explanation then acquire classification and explanation knowledge and adjust so that NewCase is correctly classified and explained.
UNTIL the expert approves the classification and explanation.
Conclusion
• Analogical problem solving technique– Supervised machine learning– Helps in reducing search space for some kind of
problems where analogy can be drawn
Case Based Reasoning– CBR gives a new way to solve problems. – It emphasizes on problem solving as well as learning.– Has got a potential of leading to significant
breakthrough in AI.
References Randolph M. Jones, Pat Langley, “Retrieval and Learning in Analogical
Problem Solving”, Proceedings of the Seventeenth Conference of the Cognitive Science Society, 1995, pp 466-471
RANDOLPH M. JONES, PAT LANGLEY “A CONSTRAINED ARCHITECTURE FOR LEARNING AND PROBLEM SOLVING”, Computational Intelligence, Volume 21, Number 4, 2005
Randolph M. Jones, “Problem Solving via Analogical Retrieval and Analogical Search Control”, Book “Foundations of Knowledge Acquisition: Machine Learning”, 1993
Mary L. Gick and Keith J. Holyoak, "Analogical Problem Solving“, Cognitive Psychology, 1983, pp 306-355
Agnar Aamodt and Enric Plaza, “Case-Based Reasoning: Foundational Issues, Methodological Variations, and System Approaches”,AI Communications, Vol. 7 Nr. 1, March 1994, pp 39-59.
Bareiss, R. (1988): PROTOS; a unified approach to concept representation, classification and learning.Ph.D. Dissertation, University of Texas at Austin, Dep.of Computer Sciences 1988. Technical Report AI88-83.