faculty-old.psau.edu.sa · Web viewInsert the root node into the queueWhile the queue is not empty...
57
CS 363 Artificial Intelligence Week 1 1. Comparison between Human and Computers with respect to intelligent factors Answer: Intelligent Parameters Human Computer Save data Limite d Yes Interaction (Speech, Audio, Video, Listen and so on) Yes No Reasoning Yes No Common Sense Yes No Decision Making Yes No Learning Ability Yes No 2. Define Artificial Intelligence Artificial Intelligence: It is the study of how to make computers with human Intelligence Think Human: Reasoning, Decision Making, Learning Ability Think rationally: Reason and Act Act Humanly: Do things in a better way on the movement like human beings. Act Rationally: Intelligent Agents 3. List foundations of AI? Answer: Mathematics, Philosophy, Psychology, Neuro Science , Computer engineering and Control theory. 4. Write the history of Artificial Intelligence? The gestation of artificial intelligence (1943–1955)
faculty-old.psau.edu.sa · Web viewInsert the root node into the queueWhile the queue is not empty Dequeue the element with the highest priority (If priorities are same, alphabetically
CS 363 Artificial Intelligence Week 1
1. Comparison between Human and Computers with respect to
intelligent factors
Answer:
Yes
No
Reasoning
Yes
No
2. Define Artificial Intelligence
Artificial Intelligence: It is the study of how to make computers
with human Intelligence
Think Human: Reasoning, Decision Making, Learning Ability
Think rationally: Reason and Act
Act Humanly: Do things in a better way on the movement like human
beings.
Act Rationally: Intelligent Agents
Answer: Mathematics, Philosophy, Psychology, Neuro Science ,
Computer engineering and Control theory.
4. Write the history of Artificial Intelligence?
The gestation of artificial intelligence (1943–1955)
The birth of artificial intelligence (1956)
Early enthusiasm, great expectations (1952–1969)
A dose of reality (1966–1973)
Knowledge-based systems: The key to power? (1969–1979)
AI becomes an industry (1980–present)
5. Does AI have applications?
Answer:
· Beating Gary Kasparov in a chess match
· Steering a driver-less car
· Robotic assistants in surgery
· Monitoring trade in the stock market to see if insider trading is
going on
6. Write the Goals of AI
· To make computers more useful by letting them take over dangerous
or tedious tasks from human
· Understand principles of human intelligence
7. Some Advantages and Dis advantages of Artificial
Intelligence
· more powerful and more useful computers
· new and improved interfaces
The Disadvantages
· increased costs
· few experienced programmers
8. AI techniques?
Answer:
· Heuristics: heuristic function, also called simply
a heuristic, is a function that ranks alternatives
in search algorithms at each branching
· Pattern recognition : The field of pattern recognition is
concerned with the automatic discovery of regularities in data
through the use of computer algorithms and with the use of these
regularities to take actions such as classifying the data into
different categories
· Machine learning: Machine learning is an application of
artificial intelligence (AI) that provides systems the ability
to automatically learn and improve from experience without being
explicitly programmed.
Week 2 & 3: Agents and Intelligent Functions
1. Define Agent ad its functions along with its examples?
Agent:
· perceiving its environment through sensors and
· acting upon its environment through actuators
Examples:
· Humans.
· Robotics.
· Web search agent.
2. What do you mean sensors/percepts and actuators/actions of Human
and Robot Agents?
Answer: Human agent:
Nose (olfaction), neuromuscular system (proprioception)… other
organs
Percepts:
Objects in the visual field (location, textures, colors,),
Auditory streams (pitch, loudness, direction), …
Actuators: hands, legs, mouth, limbs, digits, eyes, tongue……
Actions: lift a finger, turn left, walk, run, and carry an
object,
· Robotic agent:
3. Write Agent Function?
· a = F(p)
where p is the current percept , a is the action carried out, and F
is the agent function
· F maps percepts to actions F: P ® A
where P is the set of all percepts, and A is the set of all
actions
· In general, an action may depend on all percepts observed so far,
not just the current percept, so agent function is redefined
as:
· ak = F(p0 p1 p2 …pk)
where p0 p1 p2 …pk is the sequence of percepts observed to date, ak
is the resulting action carried out
· F now maps percept sequences to actions
F: P* ® A
Answer:
· e.g., A robotic car, a camera, a PC, …
· program
· The implementation of the agent function that runs on the
physical agent architecture to produce f
5. Explain Vacuum-cleaner world example by Intelligent Agent?
Answer:
Environment: two locations; square A and B.Percepts: location and
content, e.g., [A, Dirty] Actions: Left, Right, Suck, NoOp
Function:
if content == Dirty then return Suck
else if R in location == A then return Right
else if R in location == B then return Left
else if content==empty then return NoOp
6. Write about Rational Agents?
Answer:
· The rationality of an agent depends on 4 things:
The performance measure defining the agent’s degree of
success
The percept sequence, the sequence of all the things perceived by
the agent
The agent’s knowledge of the environment.The actions that the agent
can perform
· A rational agent, for each possible percept sequence, should do
an action that maximizes its performance measure based on: the
percept sequence and its built-in and acquired knowledge.
7. Rational agent design
To design an intelligent agent we need to specify the task
environment ( PEAS) as fully as possible :
PEAS: Performance measure : the goal of agent
· Environment : context, restrictions
Performance measure: Safe, fast, legal, comfortable trip, maximize
profits
• Environment: Roads, other traffic, pedestrians, customers
• Actuators: Steering wheel, accelerator, brake, signal, horn
• Sensors: Cameras, sonar, speedometer, GPS, odometer,
engine, sensors, keyboard
• Environment: Patient, hospital, staff
treatments, referrals)
answers)
Answer:
Fully observable/partially: the agent’s sensors detect all aspects
relevant to the choice of action.
Deterministic/stochastic: the next state is completely determined
by the current state and the action selected by the agent;
Episodic/sequential: the agent’s experience is divided into
“episodes”; the
Quality of the agent’s actions does not depend on previous
Episodes;
Static/dynamic: the environment remains unchanged while the agent
is choosing an action;
Discrete/continuous: there are a limited number of distinct,
clearly defined
Percepts and actions;
11. Agent types
Answer:
· Problem solving is a process of generating solutions from
observed or given data.
· A problem is characterized by: set of objects, set of states and
set of rules (actions).
· A state is an abstract representation of the agent’s environment,
OR combination of objects for agent configuration. It is specified
by the values of all objects of interest in the problem.
· Most real world problems are solved by searching for the solution
in the problem space.
· Problem space is set of all possible states generated by applying
the possible actions on the states. It may contain one or more
solutions.
· Solution is a sequence of actions from the start state to goal
state.
2. Write the steps for problem solving agents?
Answer:
· Problem formulation:
· The agent is in one of two locations, each of which
may or not contain dirt →8 possible states
· state? Robot locations: A,B – Content (dirt, clean).
· initial state? Any state
· path cost? 1unit per action
· State space has 2*2*2=8 states.
4. Example 8-Puzzle problem
Problem formulation of 8-Puzzle:
· state? locations of tiles
· Initial state? any configuration for the tiles in the
puzzle.
Integer sequences like: <7, 2, 4, 5, 0, 6, 8, 3, 1> as in the
figure given
· actions? move blank left, right, up, down
Not all moves always available
· goal test? Does the current state = goal state (given)
sequence <0, 1, 2, 3, 4, 5, 6, 7, 8>
· path cost? 1 unit per move
· 8-puzzle has 362,800 states
State Space Search for 8-Puzzle problem formulation
5. Example: 8-queens
· Formulation #1:
· states: any arrangement of 0-8 queens on the board is a
state
· Initial state: no queens on the board
· actions: add a queen to any empty square
· goal test: 8 queens are on the board, no queen attacks
another.
· Path cost: 0 (we are only interested in the solution).
64.63...57 = 1.8x1014 possible states.
6. Real world problems
7. Traveling Sales person problem
· Suppose a salesman has five cities to visit and then must return
home.
· The goal of the problem is to find the shortest path for the
salesperson to travel.find the shortest tour that visits all cities
without visiting any city twice and return to starting point.
Formulation:
· actions? move from city 1 to city 2
· goal test? a complete tour
· path cost? sum of distances
The solution: By searching through states to find a path/sequence
of all cities.
8. A Water Jug Problem
· You have a 4-gallon and
a 3-gallon water jug
amount of water
to get exactly: 2 gallons in 4-gallon jug
A Water Jug: Problem formulation:
· State representation: (x, y)
· x: Contents of 4-gallon
· y: Contents of 3-gallon
· Start state: (0, 0)
· Goal state: (2, n)
· Empty the 4-gallon jug
· Empty the 3-gallon jug
· Pour water from the 4-gallon jug into the 3-gallon jug until the
3-gallon jug is full
· Pour water from the 3-gallon jug into the 4-gallon jug until the
4-gallon jug is full
State space of Water Jug Problem:
· From the initial state, produce all successive states step by
step search tree.
9. Search tree
· Node: state in the state space
· Parent-Node: Predecessor nodes
· Action: The operator that generated the node
· Depth: number of steps along the path from the initial
state
· Path Cost: Cost of the path from the initial state to the
node
10. Evaluation criteria for Search Strategies
· Optimality: Does the strategy find the best solution (with the
lowest path cost)?
· Time and space complexity are measured in terms of:
· b:Maximum branching factor of the search tree
· d: Depth of the least-cost solution
· m: Maximum depth of the state space (may be ∞)
11. Two types of Search Techniques:
· Uninformed search (blind search)
· Informed search (heuristic search)
12.DATA STRUCTURE - BREADTH FIRST TRAVERSAL
Breadth First Search BFSBFS algorithm traverses a graph
in a breadth ward motion and uses a queue to remember to get the
next vertex to start a search, when a dead end occurs in any
iteration.
As in the example given above, BFS algorithm traverses from A to B
to E to F first then to C and G lastly to D. It employs the
following rules.
· Rule 1 − Visit the adjacent unvisited vertex. Mark it as
visited. Display it. Insert it in a queue.
· Rule 2 − If no adjacent vertex is found, remove the first
vertex from the queue.
· Rule 3 − Repeat Rule 1 and Rule 2 until the queue is
empty.
Step
Traversal
Description
1
2
We start from visiting S startingnodestartingnode, and
mark it as visited.
3
We then see an unvisited adjacent node from S. In this
example, we have three nodes but alphabetically we choose A,
mark it as visited and enqueue it.
4
Next, the unvisited adjacent node from S is B. We
mark it as visited and enqueue it.
5
Next, the unvisited adjacent node from S is C. We
mark it as visited and enqueue it.
6
Now, S is left with no unvisited adjacent nodes. So, we
dequeue and find A.
7
From A we have D as unvisited adjacent node. We
mark it as visited and enqueue it.
Algorithm:
· enqueue(obj) inserts an object into the queue.
· dequeue() removes from the queue the object that has been in
it the longest, returning this object.
· isEmpty() returns true if the queue currently
contains no objects, and false if the queue contains at
least one object.
13. Uniformed Cost Search: Uniform-cost Search: Expand node with
smallest path cost g(n).
Implementation: fringe = queue ordered by path cost
14 Depth First search Algorithm:
Depth First Search (DFS) algorithm traverses a graph in a depth
ward motion and uses a stack to remember to get the next vertex to
start a search, when a dead end occurs in any iteration.
As in the example given above, DFS algorithm traverses from S to A
to D to G to E to B first, then to F and lastly to C. It employs
the following rules.
· Rule 1 − Visit the adjacent unvisited vertex. Mark it as
visited. Display it. Push it in a stack.
· Rule 2 − If no adjacent vertex is found, pop up a vertex
from the stack. (It will pop up all the vertices from the stack,
which do not have adjacent vertices.)
· Rule 3 − Repeat Rule 1 and Rule 2 until the stack is
empty.
Step
Traversal
Description
1
2
Mark S as visited and put it onto the stack. Explore any
unvisited adjacent node from S. We have three nodes and we can
pick any of them. For this example, we shall take the node in an
alphabetical order.
3
Mark A as visited and put it onto the stack. Explore any
unvisited adjacent node from A. Both Sand D are
adjacent to A but we are concerned for unvisited nodes
only.
4
Visit D and mark it as visited and put onto the stack.
Here, we have B and C nodes, which are adjacent
to D and both are unvisited. However, we shall again
choose in an alphabetical order.
5
We choose B, mark it as visited and put onto the stack.
Here Bdoes not have any unvisited adjacent node. So, we
pop Bfrom the stack.
6
We check the stack top for return to the previous node and check if
it has any unvisited nodes. Here, we find D to be on the
top of the stack.
7
Only unvisited adjacent node is
from D is C now. So we visit C, mark it as
visited and put it onto the stack.
As C does not have any unvisited adjacent node so we keep
popping the stack until we find a node that has an unvisited
adjacent node. In this case, there's none and we keep popping until
the stack is empty.
· Step 1: SET STATUS = 1 (ready state) for each node in
G
· Step 2: Push the starting node A on the stack and set its
STATUS = 2 (waiting state)
· Step 3: Repeat Steps 4 and 5 until STACK is empty
· Step 4: Pop the top node N. Process it and set its STATUS =
3 (processed state)
· Step 5: Push on the stack all the neighbours of N that are
in the ready state (whose STATUS = 1) and set their STATUS = 2
(waiting state) [END OF LOOP]
· Step 6: EXIT
· Heuristics
1. Heuristic Search Algorithm:
Types of problem it can be applied to: Find 1 optimal solution
(when optimum value is known) Find a “close to” optimal solution
(the best solution we manage). Heuristics methods we will study:
Hill-climbing, Simulated annealing, and Genetic algorithms.
2. Characteristics of heuristic search:
· The state space is not fully explored.
· Randomization is often employed.
· There is a concept of neighborhood search.
· Heuristics are applied to explore the solutions.
· The word “heuristics” means “serving or helping to find or
discover” or “proceeding by trial and error”.
3. General Framework for Heuristic search
--Generic Optimization Problem (maximization):
Instance: A finite set X . an objective function P : X → Z. m
feasibility functions gj : X → Z, 1 ≤ j ≤ m.
Find: the maximum value of P(X) subject to X ∈ X and gj (X) ≥ 0,
for 1 ≤ j ≤ m
--Define a neighborhood function N : X → 2 X . E.g. N(X) = {X1, X2,
X3, X4, X5}. 2 Design a neighborhood search: Algorithm that finds a
feasible solution on the neighborhood of a feasible solution X.
There are two types of neighborhood searches:
I Exhaustive (chooses best profit among neighbor points)
I Randomized (picks a random point among the neighbor points)
4. Heuristic Search Algorithm:
Given N, a neighborhood function, the heuristic algorithm hN
either:
Perform one neighborhood search (using one of the strategies)
Perform a sequence of j neighborhood searches, where each one takes
us from Xi to Xi+1: [X = X0, X1, . . . , Xj = Y ].
Algorithm GenericHeuristicSearch(cmax)
Xbest ← X; (stores best so far); c ← 0;
while (c ≤ cmax) do Y ← hN (X); if (Y 6= “fail”) then X ← Y ; if
(P(X) > P(Xbest)) then Xbest ← X;
else c ← cmax + 1; (add this if hN is not randomized)]
c ← c + 1;
Stop when stuck. Problem:
It can get stuck in a local optimum.
Improvement: run the algorithm many times from different random
starting points X.
For Hill-Climbing, hN (X) returns: I Y ∈ N(X) such that Y is
feasible and P(Y ) > P(X), I or, otherwise, “fail”.
6. A* Algorithm
Insert the root node into the queue While the queue is not empty
Dequeue the element with the
highest priority (If priorities
are same, alphabetically smaller path is chosen)
If the path is ending in the
goal state, print the path and exit
Else
Insert
all the children of the dequeued element, with f(n) as the
priority
7. Greedy Best First Search Algorithm
Best-first search is a search algorithm which explores a
graph by expanding the most promising node chosen according to
a specified rule.
Week 6: Genetic Algorithms
1. A genetic algorithm (or GA) is a search technique used in
computing to find true or approximate solutions to optimization and
search problems.
· Genetic algorithms are categorized as global search
heuristics.
· Genetic algorithms are a particular class of evolutionary
algorithms that use techniques inspired by evolutionary biology
such as inheritance, mutation, selection, and crossover (also
called recombination).
2. Key terms used in Genetic Algorithms
· Individual - Any possible solution
· Search Space - All possible solutions to the problem
· Chromosome - Blueprint for an individual
· Trait - Possible aspect (features) of an individual
· Allele - Possible settings of trait (black, blond, etc.)
· Locus - The position of a gene on the chromosome
· Genome - Collection of all chromosomes for an individual
3. Genotype and Phenotype
· Phenotype:
4. GA Requirements
· A typical genetic algorithm requires two things to be
defined:
· a genetic representation of the solution domain, and
· a fitness function to evaluate the solution domain.
· A standard representation of the solution is as an array of bits.
Arrays of other types and structures can be used in essentially the
same way.
· The main property that makes these genetic representations
convenient is that their parts are easily aligned due to their
fixed size that facilitates simple crossover operation.
· Variable length representations may also be used, but crossover
implementation is more complex in this case.
· Tree-like representations are explored in Genetic
programming.
5. GA Representation
Chromosomes could be:
· Permutations of element (E11 E3 E7 ... E1 E15)
· Lists of rules (R1 R2 R3 ... R22 R23)
· Program elements (genetic programming)
· ... any data structure ...
6. GA Solution
· The fitness function is defined over the genetic representation
and measures the quality of the represented solution.
· The fitness function is always problem dependent.
· For instance, in the knapsack problem we want to maximize the
total value of objects that we can put in a knapsack of some fixed
capacity.
· A representation of a solution might be an array of bits, where
each bit represents a different object, and the value of the bit (0
or 1) represents whether or not the object is in the
knapsack.
· Not every such representation is valid, as the size of objects
may exceed the capacity of the knapsack.
· The fitness of the solution is the sum of values of all objects
in the knapsack if the representation is valid, or 0 otherwise. In
some problems, it is hard or even impossible to define the fitness
expression; in these cases, interactive genetic algorithms are
used.
7. GA Fitness Functions
Roulette Wheel’s Selection Pseudo Code:
for all members of population
sum += fitness of this individual
end for
probability = sum of probabilities + (fitness / sum)
sum of probabilities += probability
do this twice
for all members of population
if number > probability but less than next probability
then
you have been selected
· Representations
· Mutations:
· pm is called the mutation rate
· Typically between 1/pop_size and 1/ chromosome_length
· Crossovers
1. What is the Knowledge?
Answer:
· Knowledge:
· Collection of “facts” for some domain. It is a
domain-specific.
· Information about a domain that can be used to solve problems in
that domain. e.g. Computer science students are smart.
· What kinds of knowledge need to be represented in AI
systems:
1. Objects-- Facts about objects in our world
domain. e.g. Birds have wings.
2. Events-- Actions that occur in our world. e.g. Fahad
plays tennis in Abha.
3. Performance-- A behavior knowledge about how to do things. e.g.
internal processes in playing the tennis
4. Meta-knowledge-- knowledge about knowledge.
2. Explain Fundamental activities in AI about knowledge?
Answer:
· Intelligent agent should do the following activities: (KR system
tasks)
· Perceiving, acquiring knowledge from environment,
· Knowledge Representation, representing the world
information,
· Reasoning, inferring the implications(conclusions) based on what
it knows and the choices it has,
· Acting, choosing what want to do and carry out.
3. Explain Knowledge base agent?
4. Explain Types of Knowledge?
Procedural
Declarative
· Knowledge “about how to do something”.
· Focus on tasks that must be performed to reach a certain
goal.
· Represented as procedures, rules, strategies
e.g.
Determine which is older, peter or Robert, find their ages.
Find square of number….x*x
Factorial of x… x!...etc
· Focus on objects and relationships, events.
· Represented as objects, propositions, facts, logic models,
semantic nets…
e.g.
Is Ahmed is teacher?
· Knowledge representation:
-KR is to express knowledge in computer-tractable form, so that it
can be used to help AI
· agents to perform well. Thus there are two entities needed to
deal with in solving problems in AI:
Facts: Truths about the real world, this is called knowledge
level.
Representation of the facts: represent the objects in terms of
symbols that can be manipulated in programs. this is called symbol
level.
-Representation = Syntax + Semantics + Reasoning
red(car1) represents fact that my car is red.
· Syntax : How sentences are formed in the langauge?
· Semantics: the meaning of the sentences. What sentence refers to
in the real world?
· Computational aspect: how sentences and objects are represented
and manipulated to drive the conclusions.
· Suppose the language is arithmetic, then
‘x’, and ‘y’ are components of the language
The syntax says that ‘x >=y’ is a valid sentence in the
language, but ’ >>x y’ is not.
The semantics say that ‘x >= y’ is false if y is bigger than x,
and true otherwise
Week 8: Knowledge Representation and Reasoning. (PL-Logic) and
Recursion.
1. List Types of Logic systems?
Answer:
Answer:
Propositional logic sentences:
sentences in propositional logic tell you about what is true or
false.
· Each symbol P, Q, R …. is a (atomic) sentence
· Both True and False are (atomic) sentences
· A sentence wrapped in parentheses is a sentence ( )
· If P and Q are sentences, then the following are also
sentences.
· P Q conjunction
· P Q disjunction
Implication and Bi-conditional :
Conditional Proposition:
· A proposition of the form “if p then q ” or “ p implies q”,
represented “p q” is called a conditional proposition.
· “p q” : proposition p is called Hypothesis or antecedent, and the
proposition q is the Conclusion or Consequent .
· p q is false when p is true and q is false. otherwise is
true.
· e.g. “if John is from Chicago then John is from Illinois”.
2. Bi-conditional Proposition:
· The proposition p q, read “p if and only if q” ,is called
bi-conditional.
· p q is true when p and q have the same truth value, i.e., they
are both true or both false. Otherwise false.
e.g. “John is married if and only if he has a bachelor”. It is the
same as saying
“if John is married then he has a bachelor” and “if he has a
bachelor then he is married”.
Tautology, Contradiction and Satisfiability
Example:
· Prove that " It is humid today and if it is humid today then it
will rain so it will rain " is a valid argument.
Solution:
· Let us symbolize English sentences by propositional atoms as
follows:
A : It is humid today
B : It will rain
: ((A B) A) B
that is true under all four
interpretations.
A
B
Logical Equivalence
· When two compound propositions (logical expressions) S1 , S2 have
the same truth value no matter what truth value their propositions
have, they are called logically equivalent.
· Notation: S1 S2
· Read as: S1 is equivalent to S2 or S1 and S2 are logically
equivalence.
· Example:
i.e. Prove: (P⇒Q) ( ¬ P Q).
The proof is given by the truth table :
· Is the statement: (P⇒Q) (Q ⇒P). Check it!!!!!
· Is the statement: (P⇒Q) (¬ Q ⇒ ¬ P). Check it!!!!!
Equivalence rules:
Inference Properties:
· Soundness:
-An inference procedure is sound: If KB |- α then KB |= α
· Completeness:
-An inference procedure is complete: If KB |= α then KB |- α
Inference by Model Checking:
· Does KB infer α? Or
· Prove α. Can we conclude α?
· Is KB α is valid?
Inference Rules:
· Incomplete. (Why?)
Repeated application of the resolution rule to a KB in CNF may fail
to derive new valid sentences
Example: Resolution Refutation
•Note: Relations typically correspond to verbs
–Functions: Best_friend(), beginning_of() : Returns object(s)
–Connectives: , , , ,
–Quantifiers:
Represent the following sentences in first-order logic, using a
consistent vocabulary(which you must define):
a) Some students took French in spring 2001. b) Every student who
takes French passes it. c) Only one student took Greek in spring
2001. d) The best score in Greek is always higher than the best
score in French. e) Every person who buys a policy is smart. f) No
person buys an expensive policy. g) There is an agent who sells
policies only to people who are not insured.
· Student(x)Student(x): x is a student;
· French(y),Greek(y)French(y),Greek(y): y is the course
in French, resp. Greek;
· Take(x,y)Take(x,y): x(student)
takes y (course);
· Pass(x,y)Pass(x,y): x (student)
passes y(course);
· TakeInSpring2001(x,y)TakeInSpring2001(x,y): x (student)
takes y (course) in Spring 2001.
Now we can formulate the sentences as follows:
a) ∃x∃y:Student(x)∧French(y)∧TakeInSpring2001(x,y)∃x∃y:Student(x)∧French(y)∧TakeInSpring2001(x,y)
b) ∀x∀y:(Student(x)∧French(y)∧Take(x,y))Pass(x,y)∀x∀y:(Student(x)∧French(y)∧Take(x,y))Pass(x,y)
c) ∃x∃y∀z:Student(x)∧Greek(y)∧TakeInSpring2001(x,y)∧((Student(z)∧TakeInSpring2001(z,y))x=z)
All students are smart.
∃ x ( Student(x) ∧ Smart(x) )
Every student loves some other student.
∀ x ( Student(x) ⇒ ∃ y ( Student(y) ∧ ¬ (x = y) ∧ Loves(x,y)
))
There is a student who is loved by every other student.
∃ x ( Student(x) ∧ ∀ y ( Student(y) ∧ ¬(x = y) ⇒ Loves(y,x)
))
CS 363 Artificial Intelligence
Week 1
1. Comparison between Human and Computers with respect to
intelligent factors
Answer:
Yes
No
R
easoning
Yes
No
Artificial Intelligence
: It is the study of how to make computers with human
Intelligence
Think Human: Reasoning, Decision Making, Lear
ning Ability
Think rationally: Reason and Act
Act Humanly: Do things in a better way on the movement like human
beings.
Act Rationally: Intelligent Agents
Answer:
ineering and
Control theory.
–
Early enthusiasm, great expectations (1952
–
–
–
–
CS 363 Artificial Intelligence Week 1
1. Comparison between Human and Computers with respect to
intelligent factors
Answer:
Interaction (Speech, Audio, Video, Listen and so on) Yes No
Reasoning Yes No
Common Sense Yes No
Decision Making Yes No
Learning Ability Yes No
2. Define Artificial Intelligence
Artificial Intelligence: It is the study of how to make computers
with human Intelligence
Think Human: Reasoning, Decision Making, Learning Ability
Think rationally: Reason and Act
Act Humanly: Do things in a better way on the movement like human
beings.
Act Rationally: Intelligent Agents
Answer: Mathematics, Philosophy, Psychology, Neuro Science ,
Computer engineering and
Control theory.
The birth of artificial intelligence (1956)
Early enthusiasm, great expectations (1952–1969)
A dose of reality (1966–1973)
Knowledge-based systems: The key to power? (1969–1979)
AI becomes an industry (1980–present)
5. Does AI have applications?