35
Analogical Learning and Problem Solving 1 Aug 2011 Ramakrishna, 114054001 Abhijit, 114056002 Anup, 10305915

Analogical Learning and Problem Solving 1 Aug 2011 Ramakrishna, 114054001 Abhijit, 114056002 Anup, 10305915

Embed Size (px)

Citation preview

Analogical Learning and Problem Solving

1 Aug 2011

Ramakrishna, 114054001Abhijit, 114056002Anup, 10305915

Duncker's (1945) “Radiation Problem"

Attack-Dispersion story (Cont.)

Known Solution

Analogy

• The analogous solution to the radiation problem is to simultaneously direct multiple low-intensity rays toward the tumor from different directions. In this way the healthy tissue will be left unharmed, but the effects of the multiple low-intensity rays will summate and destroy the tumor

What is it ?

• Machine Learning– Supervised Learning

• Inductive Learning– Decision Trees, Version Space etc.

• Analogical Learning

– Unsupervised Learning– Reinforcement Learning

• Intuition– If two entities are similar in some respects then they

could be similar in other respects as well

Steps in Analogical Problem Solving

• Identifying Analogy– Find a known/experienced problem that is analogous to the input

problem

• Mapping– Relevant parts of the experienced problem are selected and

mapped on to the input problem to transform the new problem and derive some hypothesis for the solution

• Validation– The correctness of the newly constructed solution is validated

using theorem or simulation etc.

• Learning– If the validation is found to work well, the new knowledge is

encoded and saved for future usage.

Analogy Representation

• Appropriate level of “analogical structures” needs to be represented– Too Detailed or Too Abstract representation

will make it had to draw analogy

• Example

Analogical Structures1. A fortress was located in the center of the

country.2. Many roads radiated out from the fortress.3. A general wanted to capture the fortress

with his army.4. The general wanted to prevent mines on

the roads from destroying his army and neighboring villages.

5. As a result the entire army could not attack the fortress along one road.

6. However, the entire army was needed to capture the fortress.

7. So an attack by one small group would not succeed.

8. The general therefore divided his army into several small groups.

9. He positioned the small groups at the heads of different roads.

10. The small groups simultaneously converged on the fortress.

11. In this way the army captured the fortress.

1. A tumor was located in the interior of a patient's body.

2. -3. A doctor wanted to destroy the tumor with

rays. 4. The doctor wanted to prevent the rays from

destroying healthy tissue. 5. As a result the high-intensity rays could not

be applied to the tumor along one path.6. However, high-intensity rays were needed

to destroy the tumor. 7. So applying one low-intensity ray would not

succeed. 8. The doctor therefore divided the rays into

several low intensity rays. 9. He positioned the low intensity rays at

multiple locations around the patient's body.

10. The low-intensity rays simultaneously converged on the tumor.

11. In this way the rays destroyed the tumor.

Predicate Analogy• Sample Predicates

– locate(fortress, centre(country))– locate(tumour, interior(body))

– desire(general, capture(army, fortress))– desire(doctor, destroy(rays, tumour))

• Mapping– fortress → tumour– general → doctor– army → rays

• Inference– divide(general, small_group(army))– divide(doctor, small_group(rays))

– position(small_group(army), different_direction)– position(small_group(rays), different_direction)

– simultaneous(converge(small_group(army), fortress))– simultaneous(converge(small_group(rays),tumour))

– destroy(army, fortress)– destroy(rays, tumour)

Analogical Problem Solving

• Needs sophisticated techniques for– Representation of the problem and solution– Matching input problem with existing problem– Concluding inferences

• Very hard problem

• Let us see one system - EUREKA

EUREKA• An analogical problem solving architecture

presented by RANDOLPH M. JONES and PAT LANGLEY

• Quick look at EUREKA– Existing problem

– New problem. Solve by analogy to previous problem

• Consists of– Memory Module– Problem Solving Engine

A B C

A

B C

E F G

E

F HH

G

Memory Representation• As Semantic Network• Example network for block movement problem

Ref: Randolph Jones; Problem Solving via Analogical Retrieval and Analogical Search Control

State100

Goal17

Trans1 Transforms to a new state which satisfies the goal

Problem Solving Engine

• Uses Flexible Means-Ends Algorithm

• Two recursive functions are used– TRANSFORM – APPLY

State X State Z

Op1 State P

State C

Transform and Apply ProceduresTRANSFORM(StateX,Conditions): Returns StateZ If StateX satisfies Conditions Then Return StateZ as StateX Else Let Operator be SELEC_OPERAT0R(StateX,Conditions); If Operator is empty Then Return StateZ as "Failed State1' Else Let StateY be APPLY(StateX,Operator); If StateY is "Failed State" Then Return StateZ as "Failed State" Else Return StateZ as TRANSFORM(StateY,Conditions)

APPLY(StateX,Operator): Returns StateY Let P be PRECONDITIONS(Operator); If StateX s a t i s f i e s P Then Return StateY as EXECUTE(StateX,Operator) Else Let StateW be TRANSFORM(StateX,P); If StateW i s "Failed State" Then Return StateY as "Failed State" Else Return StateY as APPLY(StateW,Operator)

Ref: Randolph Jones; Problem Solving via Analogical Retrieval and Analogical Search Control

Example

Ref: Randolph Jones; Problem Solving via Analogical Retrieval and Analogical Search Control

• Transforming from the initial state (State100) to the final state (State200) which satisfies the given goal (Goal17)

Identifying Analogy and Problem Solving

• Collect some analogical TRANSFORM goals from the existing solution– Use spreading techniques from the current state and

goal nodes– Eg: Trans1, Trans10

• Select the best matching goals– Degree of match– History data (success/failure)

• Make concept analogy– E → A; F → B

• Proceed as per the analogical TRANSFORM goal

Identifying Analogy and Problem Solving Contd…

State500 Goal99

On5 On6

E F

On..

Closest Matching TRNSFORMATION

Goal

……….

EUREKA – Initial Stage

• EUREKA started without any knowledge in its long-term memory– For eg. First Time– never be able to solve any problems, because

there would be no previously solved problems on which to base future decisions

• Operators are stored in the form of simple problems

Case Based Reasoning (CBR) Reasoning method that uses specific past

experiences rather than a corpus of general knowledge.

Problem solving by analogy Recognizing its similarity to a specific known

problem Transferring the solution of the known problem to

the new one. Form of intra-domain analogy.

CBR terminologies Case : Denotes a problem situation Past case : A previously experienced situation which

has been captured and learned for reuse. Case based reasoning : A cyclic and integrated

process of solving a new problem.

Learning in CBR Natural by-product of problem solving. The experience of a successfully solved is retained

inorder to solve future problems If an attempt to solve problem fails, reason of failure

is remembered in order to avoid same mistake in future.

CBR Cycle RETRIEVE

Retrieve one or more previously experienced cases. REUSE

Reusing the case in one way or another. REVISE

Revising the solution based on reusing a previous case.

RETAIN Retaining the new experience by incorporating it

into existing knowledge base.

CBR Cycle

Fig . Taken from: Agnar Aamodt, Enric Plaza, “Case-Based Reasoning Foundational Issues, Methodological Variations, and System Approaches”, AI communications, 1994.

How “cases” are represented ?

Figure taken from “Agnar Aamodt and Enric Plaza, “Case-Based Reasoning: Foundational Issues, Methodological Variations, and System Approaches”,AI Communications, Vol. 7 Nr. 1, March 1994, pp 39-59”

How “cases” are represented ? In dynamic memory model, cases that share similar properties are

organized under a more general structure known as generalized episode – GE.

A generalized episode contains

Norms

denote features common to all cases indexed under a GE. Indices

are features that discriminate between a GE's cases may point to a more specific generalized episode, or directly

to a case. An index is composed of

Index name Index value

Cases

How to find a similar problem ?

When a new case description is given and the best matching is searched for, the input case structure is 'pushed down' the network structure, starting at the root node.

When one or more features of the case match one or more features of a GE, the case is further discriminated based on its remaining features.

Eventually, the case with most features in common with the input case is found.

How a new case is trained to system ?

During storing of a new case, it is discriminated by indexing it under different indices below its most specific generalized episode.

If, during the storage of a case - two cases (or two GEs) end up under the same index, a new generalized episode is automatically created.

An example...

Generalized Episode 2

Norms: Fever

Indices:

1 2 3

Influenza symptoms

Swine Flu Symptoms

Malaria symptoms

Values:

Index:

Influenza

Swine Flu

MalariaCases:

Generalized Episode 3

Norms: Dysentry

Indices:

1 2 3

food poisoning symptoms

Cholera Symptoms

Amoebicsymptoms

Values:

Index:

food poisoning

Cholera Amoebic dysentry

Cases:

Generalized Episode 1

Norms: Disease symptoms

Indices:

1 2

Fever symptoms

Dysentry symptoms

Values:

Index:

An example... Consider a doctor patient senario.

The case of diseases trained are already stored in the tree fashion.

It has got Generalized Episodes(GE) as the entity.

Suppose a new patient comes with malaria symptoms.

Current case symptoms are compared with the already existing ones. So it is directed to the left subtree. Here norms give the features common to that GE.

Then cases are distinguished according to the symptoms which are specific to those.

Ultimately the current case is directed to the index 3 where it is matched and the solution to the already stored case is applied to the current case.

Protos Case-based problem solving and learning system for

heuristic classification tasks.

In Protos, a concept ci is represented extensionally as a collection of examples (called exemplars or cases):

ci = {ei1, ei2, ...}.

The classification and learning algorithmInput: a set of exemplar-based categories C = {c1, c2, ... , cn} and a case (NewCase)

to classify.

REPEAT

Classify:

Find an exemplar of ci є C that strongly matches NewCase and classify NewCase as ci.

Explain the classification.

Learn:

If the expert disagrees with the classification or explanation then acquire classification and explanation knowledge and adjust so that NewCase is correctly classified and explained.

UNTIL the expert approves the classification and explanation.

Conclusion

• Analogical problem solving technique– Supervised machine learning– Helps in reducing search space for some kind of

problems where analogy can be drawn

Case Based Reasoning– CBR gives a new way to solve problems. – It emphasizes on problem solving as well as learning.– Has got a potential of leading to significant

breakthrough in AI.

References Randolph M. Jones, Pat Langley, “Retrieval and Learning in Analogical

Problem Solving”, Proceedings of the Seventeenth Conference of the Cognitive Science Society, 1995, pp 466-471

RANDOLPH M. JONES, PAT LANGLEY “A CONSTRAINED ARCHITECTURE FOR LEARNING AND PROBLEM SOLVING”, Computational Intelligence, Volume 21, Number 4, 2005

Randolph M. Jones, “Problem Solving via Analogical Retrieval and Analogical Search Control”, Book “Foundations of Knowledge Acquisition: Machine Learning”, 1993

Mary L. Gick and Keith J. Holyoak, "Analogical Problem Solving“, Cognitive Psychology, 1983, pp 306-355

Agnar Aamodt and Enric Plaza, “Case-Based Reasoning: Foundational Issues, Methodological Variations, and System Approaches”,AI Communications, Vol. 7 Nr. 1, March 1994, pp 39-59.

Bareiss, R. (1988): PROTOS; a unified approach to concept representation, classification and learning.Ph.D. Dissertation, University of Texas at Austin, Dep.of Computer Sciences 1988. Technical Report AI88-83.