Soft Computing PDF

Embed Size (px)

Citation preview

  • 8/13/2019 Soft Computing PDF

    1/103

    SOFT COMPUTING

  • 8/13/2019 Soft Computing PDF

    2/103

    2

    Human sciences

    quantitative qualitative

    numerical dataprecise objects

    conventional logic

    complicated mathematicscomputer models

    non-numerical dataimprecise objects

    approximate reasoning

    interpretationmanual work

    methodsare

    methodsare

    that means that means

    often usedseparately

    social sciencesbehavioral sciences

    the humanitieseconomics

    law

    medicine

    include

  • 8/13/2019 Soft Computing PDF

    3/103

    3

    Traditional Approaches to Computerized Modeling

    Mathematical models:Complicated, black

    boxes, numbercrunching.

    Rule-based systems(crisp & bivalent):

    Large rule bases.

  • 8/13/2019 Soft Computing PDF

    4/103

    What is Hard Computing?

    Hard computing, i.e., conventional computing, requires a

    precisely stated analytical model and often a lot of

    computation time.

    Many analytical models are valid for ideal cases.

    Real world problems exist in a non-ideal environment.

    Premises and guiding principles of Hard Computing are

    Precision, Certainty, and rigor. Many contemporary problems do not lend themselves to precise

    solutions such asRecognition problems (handwriting, speech, objects,

    images)

    Mobile robot coordination, forecasting, combinatorial problems etc.

  • 8/13/2019 Soft Computing PDF

    5/103

    M

    V

    L

    s(t) H(t)

    md2/dt2(s(t)+Lsin(t) = H(t)md2/dt2(Lcos(t)) = V(t)-mgJd2/dt2= (LV(t)sin(t)-LH(t)cos(t)) = V(t)-mgMd2/dt2s(t) = (t)-H(t)-Fd/dts(t)

  • 8/13/2019 Soft Computing PDF

    6/103

    Soft Computing (SC)

    mimic humanreasoning that

    is oftenlinguistic and

    approximate bynature

    fuzzy systems, probabilistic reasoning,natural computing (evolutionary computing,

    cellular automata, DNA computing,neurocomputing, immunocomputing,

    swarm theory, etc.).

    imprecisionlearning

    uncertaintyoptimization

    adaptive andintelligent systems

    orcomputational

    intelligence

    mathematical orstatistical methods(hybrid methods)

    aims to includes

    also known as

    can copewith

    can be usedwith

  • 8/13/2019 Soft Computing PDF

    7/103

    What is Soft Computing ?

    (adapted from L.A. Zadeh)

    Soft computing differs from conventional (hard) computingin that, unlike hard computing, it is tolerant of imprecision,

    uncertainty, partial truth, and approximation. In effect, the role

    model for soft computing is the human mind.

    Guiding Principles of Soft Computing The guiding principle of soft computing is:

    Exploit the tolerance for imprecision, uncertainty, partial truth, and approximation

    to achieve tractability, robustness and low solution cost.

  • 8/13/2019 Soft Computing PDF

    8/103

    As complexi ty r ises, precise

    statements lose meaning and

    meaningfu l statements loseprecision

    -

    Lotfi Zadeh

  • 8/13/2019 Soft Computing PDF

    9/103

    Premises of Soft Computing

    The real world problems are pervasively imprecise and uncertain

    Precision and certainty carry a cost

  • 8/13/2019 Soft Computing PDF

    10/103

    Techniques in Soft ComputingNeural networks

    Fuzzy Logic

    Genetic Algorithms in Evolutionary ComputationParticle swarm optimization

    Neuro-Fuzzy systems

  • 8/13/2019 Soft Computing PDF

    11/103

    Unique Property of Soft computing Learning from experimental data

    Soft computing techniques derive their power of generalization from

    approximating or interpolating to produce outputs from previously unseeninputs by using outputs from previous learned inputs

    Generalization is usually done in a high dimensional space.

  • 8/13/2019 Soft Computing PDF

    12/103

    Current Applications using Soft Computing Application of soft computing to handwriting recognition

    Application of soft computing to automotive systems and

    manufacturing

    Application of soft computing to image processing and

    data compression Application of soft computing to architecture

    Application of soft computing to decision-support

    systems

    Application of soft computing to power systems

    Neurofuzzy systems

    Fuzzy logic contro

  • 8/13/2019 Soft Computing PDF

    13/103

    Future of Soft Computing(adapted from L.A. Zadeh)

    Soft computing is likely to play an especially important role in science

    and engineering, but eventually its influence may extend much

    farther. Soft computing represents a significant paradigm shift in the aims of

    computing

    a shift which reflects the fact that the human mind,

    unlike present day computers, possesses a remarkable ability to store

    and process information which is pervasively imprecise, uncertain and

    lacking in categoricity.

  • 8/13/2019 Soft Computing PDF

    14/103

    14

    SC Applications: Control

    Heavy industry(Matsushita, Siemens,Stora-Enso, Metso)

    Home appliances (Canon,Sony, Goldstar, Siemens)

    Automobiles (Nissan,Mitsubishi, Daimler-Chrysler, BMW,Volkswagen)

    Space crafts (NASA)

    http://web1.mitsubishi-motors.co.jp/inter/technology/technology.html
  • 8/13/2019 Soft Computing PDF

    15/103

    Fuzzy logic in Reality (Industrial Applications)

    Efficient and stable control of car-engines

    (Nissan)Simplified control of robots (Hirota,Fuji Electric, Toshiba, Omron)

    Industrial control applications

    (Aptronix, Omron, Meiden, Micom,

    Mitsubishi, Nissin-Denki, Oku-Electronics)Archiving system for documents

    (Mitsubishi Elec.)

    Prediction system for early recognitionof earthquakes

    (Bureau of Metrology, Japan)

    Recognition of handwritten symbols withpocket computers (Sony)

    Video cameras (Canon, Minolta)

    Washing-machines (Matsushita,Hitatchi, Samsung)

    Recognition of handwriting,objects, voice (Hitachi, HosaiUniv., Ricoh)

    Efficiency for elevator control(Fujitec, Hitachi, Toshiba)

    Positioning of wafer-steppers inthe production ofsemiconductors

    (Canon)

    Automatic control of dam gatesfor hydroelectric-power plants(Tokyo Electric Power)

  • 8/13/2019 Soft Computing PDF

    16/103

    Components of soft computing include:Neural networks(NN)Fuzzy logics(FL)Evolutionary computation(EC),

    including:Evolutionary algorithms

    Genetic algorithmsDifferential evolution

    Metaheuristicand SwarmIntelligence

    Ant colony optimizationBees algorithmsBat algorithmCuckoo searchHarmony searchFirefly algorithm

    Artificial immune systemsParticle swarm optimization

    Ideas about probabilityincluding:Bayesian network

    Chaos theoryPerceptron

    http://en.wikipedia.org/wiki/Neural_networkhttp://en.wikipedia.org/wiki/Fuzzy_logichttp://en.wikipedia.org/wiki/Evolutionary_computationhttp://en.wikipedia.org/wiki/Evolutionary_algorithmhttp://en.wikipedia.org/wiki/Genetic_algorithmhttp://en.wikipedia.org/wiki/Differential_evolutionhttp://en.wikipedia.org/wiki/Metaheuristichttp://en.wikipedia.org/wiki/Swarm_Intelligencehttp://en.wikipedia.org/wiki/Swarm_Intelligencehttp://en.wikipedia.org/wiki/Ant_colony_optimizationhttp://en.wikipedia.org/wiki/Bees_algorithmhttp://en.wikipedia.org/wiki/Bat_algorithmhttp://en.wikipedia.org/wiki/Cuckoo_searchhttp://en.wikipedia.org/wiki/Harmony_searchhttp://en.wikipedia.org/wiki/Firefly_algorithmhttp://en.wikipedia.org/wiki/Artificial_immune_systemhttp://en.wikipedia.org/wiki/Particle_swarm_optimizationhttp://en.wikipedia.org/wiki/Probabilityhttp://en.wikipedia.org/wiki/Bayesian_networkhttp://en.wikipedia.org/wiki/Chaos_theoryhttp://en.wikipedia.org/wiki/Perceptronhttp://en.wikipedia.org/wiki/Perceptronhttp://en.wikipedia.org/wiki/Chaos_theoryhttp://en.wikipedia.org/wiki/Bayesian_networkhttp://en.wikipedia.org/wiki/Probabilityhttp://en.wikipedia.org/wiki/Particle_swarm_optimizationhttp://en.wikipedia.org/wiki/Artificial_immune_systemhttp://en.wikipedia.org/wiki/Firefly_algorithmhttp://en.wikipedia.org/wiki/Harmony_searchhttp://en.wikipedia.org/wiki/Cuckoo_searchhttp://en.wikipedia.org/wiki/Bat_algorithmhttp://en.wikipedia.org/wiki/Bees_algorithmhttp://en.wikipedia.org/wiki/Ant_colony_optimizationhttp://en.wikipedia.org/wiki/Swarm_Intelligencehttp://en.wikipedia.org/wiki/Swarm_Intelligencehttp://en.wikipedia.org/wiki/Metaheuristichttp://en.wikipedia.org/wiki/Differential_evolutionhttp://en.wikipedia.org/wiki/Genetic_algorithmhttp://en.wikipedia.org/wiki/Evolutionary_algorithmhttp://en.wikipedia.org/wiki/Evolutionary_computationhttp://en.wikipedia.org/wiki/Fuzzy_logichttp://en.wikipedia.org/wiki/Neural_network
  • 8/13/2019 Soft Computing PDF

    17/103

    What Is A Neural Network?

    The simplest definition of a neural network, more properlyreferred to as an 'artificialneural network (ANN), is providedby the inventor of one of the first neurocomputers, Dr.Robert Hecht-Nielsen.

    He defines a neural network as: "...a computing system

    made up of a number of simple, highly interconnectedprocessing elements, which process information by theirdynamic state response to external inputs.

    In "Neural Network Primer: Part I" by Maureen Caudill, AI

    Expert, Feb. 1989

  • 8/13/2019 Soft Computing PDF

    18/103

    Neural net work definations

    In general form, a neural network is a machinethat is designed to model the way in which the

    brain performs a particular task or function of

    interest

    a neural network is a system composed of many

    simple processing elements operating in parallel

    whose function is determined by network

    structure, connection strengths, and theprocessing performed at computing elements

    or nodes

  • 8/13/2019 Soft Computing PDF

    19/103

    A neural network is a massively parallel distributed

    processor made up of simple processing units, whichhas a natural propensity for storing experiential

    knowledge and making it available for use.

    It resembles the brain in two respects:

    Knowledge is acquired by the network through a

    learning process.

    Interneuron connection strengths known as synaptic

    weights are used to store the knowledge

  • 8/13/2019 Soft Computing PDF

    20/103

    Applications

    Non linearity.

    A neural network, made up of aninterconnection of nonlinear neurons is

    itself nonlinear.

    Input-output mapping

  • 8/13/2019 Soft Computing PDF

    21/103

    Adaptivity; Neural networks have a builtin capacity to adapt their synaptic weightsto changes in the surrounding

    environment.

    Evidential response: Indicates the

    confidence in decision made;

  • 8/13/2019 Soft Computing PDF

    22/103

    McCulloch and Pitts produced the first

    neural network in 1943Many of the principles can still be seen

    in neural networks of today

  • 8/13/2019 Soft Computing PDF

    23/103

    If the brain were so

    simple that we could

    understand it then wedbe so simple that we

    couldntLyall Watson

  • 8/13/2019 Soft Computing PDF

    24/103

    Biological inspiration

    Animals are able to react adaptively to changes in their external andinternal environment, and they use their nervous system to performthese behaviours.

    An appropriate model/simulation of the nervous system should beable to produce similar responses and behaviours in artificialsystems.

    The nervous system is build by relatively simple units, the neurons,so copying their behavior and functionality should be the solution.

  • 8/13/2019 Soft Computing PDF

    25/103

    Biological inspiration

    Dendrites

    Soma (cell body)

    Axon

  • 8/13/2019 Soft Computing PDF

    26/103

    Biological inspiration

    synapses

    axondendrites

    The information transmission happens at the synapses.

  • 8/13/2019 Soft Computing PDF

    27/103

    We are born with about 100 billion neurons

    A neuron may connect to as many as 100,000 other neurons

  • 8/13/2019 Soft Computing PDF

    28/103

    Signals move via electrochemical

    signals The synapses release a chemical

    transmitterthe sum of which can

    cause a threshold to be reachedcausing the neuron to fire

    Synapses can be inhibitory or

    excitatory

  • 8/13/2019 Soft Computing PDF

    29/103

    Comparison of Brains and Traditional

    Computers

    200 billion neurons, 32

    trillion synapses Element size: 10-6m

    Energy use: 25W

    Processing speed: 100 Hz

    Parallel, Distributed

    Fault Tolerant

    Learns: Yes

    Intelligent/Conscious:Usually

    1 billion bytes RAM but

    trillions of bytes on disk Element size: 10-9 m

    Energy watt: 30-90W (CPU)

    Processing speed: 109 Hz

    Serial, Centralized

    Generally not Fault Tolerant

    Learns: Some

    Intelligent/Conscious:Generally No

  • 8/13/2019 Soft Computing PDF

    30/103

    Neurons in the Brain

    Although heterogeneous, at a low levelthe brain is composed of neurons A neuron receives input from other neurons

    (generally thousands) from its synapses

    Inputs are approximately summed

    When the input exceeds a threshold theneuron sends an electrical spike that travelsthat travels from the body, down the axon, tothe next neuron(s)

  • 8/13/2019 Soft Computing PDF

    31/103

    Biological inspiration

    The spikes travelling along the axon of the pre-synaptic neurontrigger the release of neurotransmitter substances at the synapse.

    The neurotransmitters cause excitation or inhibition in the

    dendrite of the post-synaptic neuron.The integration of the excitatory and inhibitory signals mayproduce spikes in the post-synaptic neuron.

    The contribution of the signals depends on the strength of thesynaptic connection.

  • 8/13/2019 Soft Computing PDF

    32/103

    McCulloch & Pitts (1943) are generally recognised as thedesigners of the first neural network

    Many of their ideas still used today (e.g. many simple units

    combine to give increased computational power and the idea of athreshold

    Hebb (1949) developed the first learning rule (on the premisethat if two neurons were active at the same time the strength

    between them should be increased

    During the 50s and 60s many researchers worked on theperceptron amidst great excitement.

    1969 saw the death of neural network research for about 15

    yearsMinsky & Papert

    Only in the mid 80s(Parker and LeCun) was interest revived (infact Werbos discovered algorithm in 1974

  • 8/13/2019 Soft Computing PDF

    33/103

  • 8/13/2019 Soft Computing PDF

    34/103

    We are born with about 100 billion neurons

    A neuron may connect to as many as 100,000 other neurons

  • 8/13/2019 Soft Computing PDF

    35/103

    Fault tolerance a neural net workimplemented in hardware form, has thepotential to be inherently fault tolerent,or

    capable of robust computation VLSI Implement- ability. The massive

    parallel nature of a neural network makes

    it potentially fast for the computation ofcertain task.

  • 8/13/2019 Soft Computing PDF

    36/103

  • 8/13/2019 Soft Computing PDF

    37/103

  • 8/13/2019 Soft Computing PDF

    38/103

  • 8/13/2019 Soft Computing PDF

    39/103

  • 8/13/2019 Soft Computing PDF

    40/103

  • 8/13/2019 Soft Computing PDF

    41/103

  • 8/13/2019 Soft Computing PDF

    42/103

  • 8/13/2019 Soft Computing PDF

    43/103

  • 8/13/2019 Soft Computing PDF

    44/103

  • 8/13/2019 Soft Computing PDF

    45/103

  • 8/13/2019 Soft Computing PDF

    46/103

    http://en.wikipedia.org/wiki/File:Neural_network_example.svg
  • 8/13/2019 Soft Computing PDF

    47/103

    Simplified view of a feedforward artificial neural network

    http://en.wikipedia.org/wiki/File:Neural_network_example.svghttp://en.wikipedia.org/wiki/File:Neural_network_example.svghttp://en.wikipedia.org/wiki/File:Neural_network_example.svghttp://en.wikipedia.org/wiki/File:Neural_network_example.svg
  • 8/13/2019 Soft Computing PDF

    48/103

  • 8/13/2019 Soft Computing PDF

    49/103

  • 8/13/2019 Soft Computing PDF

    50/103

    Network architecture

    Feed forward network

    60 input (one for each frequency bin)

    6 hidden

    2 output (0-1 for Steve, 1-0 for David)

  • 8/13/2019 Soft Computing PDF

    51/103

    Presenting the dataSteve

    David

  • 8/13/2019 Soft Computing PDF

    52/103

    What is meant by learning in NN ?

    Learning is a process by which the freeparameters of a neural network are adapted

    through a process of stimulation by the

    environment in which the network is

    embedded.

    The type of lerning is determined by the

    manner in which the parameters changes takeplace.

  • 8/13/2019 Soft Computing PDF

    53/103

    Classification of Learning algorithms

    Supervised Learning:Error Correction, Gradient descent :

    LMS , Back PropStochastic (Boltzmann)

    Re-inforcement learning

    Unsupervised learningHebbianCompetitive

    S i d l i A i ti l i i hi h th t k i

  • 8/13/2019 Soft Computing PDF

    54/103

    Supervised learning or Associative learning in which the network istrained by providing it with input and matching output patterns.These input-output pairs can be provided by an external teacher

  • 8/13/2019 Soft Computing PDF

    55/103

    Reinforcement Learning / Neurodynamic

    Programming

    In supervised learning we have assumed that there is a target output value for each

  • 8/13/2019 Soft Computing PDF

    56/103

    In supervised learning we have assumed that there is a target output value for eachinput value. However, in many situations, there is less detailed information available.In extreme situations, there is only a single bit of information after a long sequenceof inputs telling whether the output is right or wrong. Reinforcement learning isone method developed to deal with such situations.

    Reinforcement learning (RL) is a kind of supervised learning in that some feedbackfrom the environment is given. However the feedback signal is only evaluative, notinstructive. Reinforcement learning is often called learning with a criticas opposedto learning with a teacher.

    Humans learn by interacting with the environment. When a baby plays, it waves itsarms around, touches things, tastes things, etc. There is no explicit teacher but

    there is a sensori-motor connection to its environment. Such a connectionprovides information about cause and effect, the consequence of actions, and whatto do to achieve goals.

    Learning from interaction with our environment is a fundamental idea underlyingmost theories of learning. RL has rich roots in the psychology of animal learning,from where it gets its name.

    The growing interest in RL comes in part from the desire to build intelligentsystems that must operate in dynamically changing real- world environments.

    Robotics is the common example.

  • 8/13/2019 Soft Computing PDF

    57/103

    Unsupervised learning

  • 8/13/2019 Soft Computing PDF

    58/103

    Error correction method

    Error correction is, as the name indicates, about correcting the

    synaptical strengths according to the error in the neurons output.

    Mathematically the definition of a neurons output error is given by

    this equation:

    This equation is one of the important ones, and is used frequentlythroughout this page. It is explained more thoroughly on the next

    page, where also the time is involved.

  • 8/13/2019 Soft Computing PDF

    59/103

  • 8/13/2019 Soft Computing PDF

    60/103

    The change of the neuron ks synaptic weight is determied by the

    error of the neurons output, the associated synapses and a learning

    rate. This rule is named The Widrow-Hoff-rule, or simply The

    Delta-rule.

    The larger error the larger change.

    The input signals (synapses) xjalso determins the rate of learning.

    If an input signal is small, it havent got much influence in the

    neurons error. The associated synaptic weight, should likewise not

    be changed much.

    The learning rate is an auxillary value determining the rate of

    learning. It can be set by the teacher to control the updates of the

    synapticweights.

  • 8/13/2019 Soft Computing PDF

    61/103

    The update of neuron k's synaptical strengths is now very

    simple. The future strength of synapse j in neuron k has to beequal to the present plus the given alteration of the synapse.

    The synapses of neuron k are now altered and the result of thesame input should now be closer to the wanted than before.

    P t L i

  • 8/13/2019 Soft Computing PDF

    62/103

    Perceptron Learning

  • 8/13/2019 Soft Computing PDF

    63/103

  • 8/13/2019 Soft Computing PDF

    64/103

  • 8/13/2019 Soft Computing PDF

    65/103

  • 8/13/2019 Soft Computing PDF

    66/103

  • 8/13/2019 Soft Computing PDF

    67/103

  • 8/13/2019 Soft Computing PDF

    68/103

  • 8/13/2019 Soft Computing PDF

    69/103

  • 8/13/2019 Soft Computing PDF

    70/103

  • 8/13/2019 Soft Computing PDF

    71/103

  • 8/13/2019 Soft Computing PDF

    72/103

  • 8/13/2019 Soft Computing PDF

    73/103

  • 8/13/2019 Soft Computing PDF

    74/103

    Competitive learning

  • 8/13/2019 Soft Computing PDF

    75/103

    Competitive learning is a simple learning algorithm, which is veryuseful for classify set of input patterns. In competitive learningthere is only one active neuron at a time, and it's only the activeneuron that is trained.

    Three basic elements of competitive learning rule

    A set of neurons that all are same except for some randomlydistribute synaptic weights which therefore respond differently togiven set of input patters

    A limit imposed on the strength of each neuron

    Mechanism that permits the neuron to compete for the right torespond to a given subset of inputs such that only one outputneuron or only one neuron group is active at a time the neuron that

    wins the competition is called a winner takes all neuron

    Competitive learning

  • 8/13/2019 Soft Computing PDF

    76/103

  • 8/13/2019 Soft Computing PDF

    77/103

    Actually this just means that all outputs is run through and the

    neuron with the largest output, has an output at 1. The rest of theneurons will have 0 as output.

    Hebbian learning

  • 8/13/2019 Soft Computing PDF

    78/103

    Donald O. Hebb presented a theory of behaviour

    based as much as possible on the physiology of the

    nervous system.

    The most important concept to emerge from Hebb's work was his formal

    statement (known as Hebb's postulate) of how learning could occur. Learning

    was based on the modification of synaptic connections between neurons.

    Specifically,

    When an axon of cell A is near enough to excite a

    cell B and repeatedly or persistently takes part in

    firing it, some growth process or metabolic change

    takes place in one or both cells such that A's

    efficiency, as one of the cells firing B, is increased.

  • 8/13/2019 Soft Computing PDF

    79/103

    .

    1. if two neurons on either side of synapse areactivate simultaneously then the strength of that

    synapse is selectively increased.

    2. If two neurons on either side of a synapse are

    activated asynchronously, the that synapse is

    selectively weakned or eliminated.

    H bbi

  • 8/13/2019 Soft Computing PDF

    80/103

    Hebbian synapse

    Time dependent mechanism. This mechanism refers to the fact that modification in a

    hebbian synapse depend on the exact time of occurrence of the presynaptic and

    post synaptic signals.

    Local mechanism. The local available information is used by a Hebbian synapse to

    produce local synaptic modification that is input specific.

    Interactive mechanism. The occurrence of a change in a Hebbian synapse depends

    on signals on both sides of synapse.(deterministic or stochastic)

    conjunctional or correlation Mechanism;Conjuction( co-ocurrence) of presynaptic and post synaptic signals with in a short

    interval of time is sufficient to produce the synaptic modification.

    And the correlation between presynaptic and post synaptic signals viewed as being

    responsible for synaptic changes.

    S M d f

  • 8/13/2019 Soft Computing PDF

    81/103

    Synaptic Modifications

    Hebbian:Hebbian synapse increases its strengthfor positively correlated presynaptic and

    postsynaptic signals and decreases its strength

    when these signals are either uncorrelated or

    negatively correlated Anti Hebbian: strengthen the negatively

    correlated signals

    Non Hebbian: Does not involve Hebbianmechanism

  • 8/13/2019 Soft Computing PDF

    82/103

    The basic idea is that if two units j and k are active simultaneously, theirinterconnection must be strengthened. If j receives input from k, thesimplest version of Hebbian learning is described by

    Application of the hebbian learning rule: the linear

  • 8/13/2019 Soft Computing PDF

    83/103

    associator

    Take a network of neurons which are organized in two distinct layers (calledf and g). Neurons in layer f project to neurons in layer g but not vice versa.Each neuron in layer g is a linear unit, its output is the sum of its inputs(hence the name linear associator). In vector notations, this is equivalent towriting: g = a*fThe strength of the connection form presynaptic neuron f[j] to postsynpatic

    neuron g[i] is given by a[i,j].This notation may seem strange (w[i,j] denotingthe connection going from neuron j to neuron i but is the most commonlyused). The activation of each neuron in the output layer is given by is sum ofweighted inputs. The strength of each connection is calculated from theproduct of the pre- and postsynaptic activities scaled by a learningrate(which determines how fast connection weights change).

    wij= * g[i] * f[j]. in vector notations, this is equivalent to writing: w= *g* fT. We will calculate anexample in class!

    The linear associator stores associations between a pattern of neural activationsin the input layer f and a pattern of activations in the output layer g Once the

  • 8/13/2019 Soft Computing PDF

    84/103

    in the input layer f and a pattern of activations in the output layer g. Once theassociations have been stored in the connection weights between layer f andlayer g, the pattern in layer g can be recalled by presentation of the inputpattern in layer f.

    An auto-associator stores association within a layer of neurons. Once theassociations between the neural activities in a given pattern are stored in theconnection weights, the auto-associator can recall the storedpattern from a noisy or incomplete input pattern.

  • 8/13/2019 Soft Computing PDF

    85/103

    Boltzmann Learning(stochastic method)

    Boltzmann learning is statistical in nature, and is derived from thefield of thermodynamics. It is similar to error-correction learningand is used during supervised training. In this algorithm, the state ofeach individual neuron, in addition to the system output, are takeninto account. In this respect, the Boltzmann learning rule issignificantly slower than the error-correction learning rule. Neuralnetworks that use Boltzmann learning are called Boltzmannmachines.

    Boltzmann learning is similar to an error-correction learning rule, inthat an error signal is used to train the system in each iteration.However, instead of a direct difference between the result valueand the desired value, we take the difference between theprobability distributions of the system.

  • 8/13/2019 Soft Computing PDF

    86/103

    Boltzmann machine properties

    Recurrent network with all nodes connected to each other

    Nodes have binary outputs (either 0,1 or -1,1) Weights between the nodes are symmetric

    Connection from a node to itself is not allowed

    Nodes are updated asynchronously (i.e. nodes are selected at random)

    The network can have hidden nodes

    Learning can be supervised or unsupervised

    Node activation is stochastic

    The machine is characterized by the energy function E and value dependsupon particular states occupied by individual neurons where xjis the stateof j neuron and no self feed back

  • 8/13/2019 Soft Computing PDF

    87/103

    Activation algorithm

    A node i is chosen at random for updating and the state of the node is setto 1 with a probability :

    where is the energy difference from the ith unit being on or off. Thisdifference is given by derivation of the energy E with respect to the nodes

    state and is therefor:

    And T is a parameter which simulates the temperature of a physicalsystem, it is used with simulated annealing in the learning process.

  • 8/13/2019 Soft Computing PDF

    88/103

    Neurons in boltzmann machine can be of two

    groups:Visible neurons and Hidden neurons There are two modes of operation: clamped

    condition and free-running condition ( for both

    neurons)

    Where P+IJ is correlation under clamped condition between i and j neuronAnd p-ij is correlation under free running condition and i not equal to k

    .

  • 8/13/2019 Soft Computing PDF

    89/103

    The network can be divided into to sets of nodes, a nonempty set ofvisible nodes V and a set of hidden nodes H. Visible nodes are those which

    are the interface towards the environment or the outside world. Thehidden nodes used only for internal representation or modeling and doesnot interact with the outside world.

    Supervised learning can be achieved by declaring some of the visible nodesas input nodes and some as output nodes. When we have no output nodesamong the visible nodes we have unsupervised learning.

    The learning algorithm has two different phases. The clamping phase andthe free-running phase. In the clamping phase we lock or clampa set ofthe visible nodes to specific values given by a specific pattern. Nodes thatare clamped wont change their output value and then we sample thevalues of the hidden nodes using simulated annealing.

    In the free-running phase no nodes are locked/clamped and we samplevalues for all the nodes using simulated annealing.

    M B d L i

  • 8/13/2019 Soft Computing PDF

    90/103

    Memory Based Learning

    Memory Based Learning (MBL) is a simple function approximationmethod whose roots go back at least to 1910. Training a memorybased learner is an almost trivial operation: just store each datapoint in memory (or a database). Making a prediction about theoutput that will result from some input attributes based on the data

    is done by looking for similar points in memory, fitting a local modelto those points, and then making a prediction based on the model.There are four components that define a memory based learner: adistance metric, the number of nearest neighbors, a weightingfunction, and a local model. Each will be described in turn in the nextfour subsections followed by a discussion of multivariate

    considerations and classification problems.

    Memory based learning

  • 8/13/2019 Soft Computing PDF

    91/103

    Memory-based learning is actually a method of storing past experience. Its storesinput/output examples in a set of vectors for later use in a large memory grouping.

    A memory-based system needs two things: criterion which defines the localneighborhood of inputs, and learning rules which are applied to the training examples inthat local neighborhood.

    Essentially the system is given a set of examples and when given a set of inputs it triesto figure out which one it most closely resembles. Because of this system's skill inclassifying, even an infinite size learning set, only half of it would ever be used.

    This type of system is exceptional at identifying outliers in a data set in order toapproximate or classify its category or function [4].

    XN { x1, x2,xN} is said to be nearest neighbor of vectror Xtestif min d(xi, xtest) = d (xN, Xtest) where d (xN, Xtest) is the Euclidean distance between the vectors xi,

    and xtest

    Neural network applications

  • 8/13/2019 Soft Computing PDF

    92/103

    Neural network applications

    Pattern association: There are two types of patternassociation, autoassociationand heteroassociation.

    Autoassociation is about constantly showing the network a certain

    pattern, the network should be able to store this pattern, and then when

    a distorted image of the same pattern shows up, the task is to retrieve

    the pattern.

    Heteroassociationis different in the sense that it is supervised, and it is

    a pairing of a set of input patterns and a set of output patterns. Mostly it

    is about the retrieving phase, xkis a key pattern and ykis a memorised

    pattern, the task is to make the network retrieve ykat the presentation of

    xk.

    Pattern

    associator

    Output vectorY

    Input vector

    Storage phase andRecall phase

    Patterniti

    Unsupervised network for

    feature extraction

    supervised network for

    classifications

  • 8/13/2019 Soft Computing PDF

    93/103

    recognition

    Pattern recognition is formally defined as the process where by a receivedpattern/ signal is assigned to one of a prescribed number of classes(categories)

    m-dimensionalobservation space

    q-dimensional featurespace

    r-dimensional

    decision space

    Input pattern x Feature vector Y

    Function Approximation

  • 8/13/2019 Soft Computing PDF

    94/103

    pp Function Approximation: The neural network can also be used to

    approximate functions, this means that the network is able to receive an

    input and a desired output, and then approximate the function that hasbeen used,

    Let d = f(x)Let's say that F(x)is the real function used, the function that thenetwork comes up with, will be very close by: F(x) - f(x) < eWhere e is asmall positive number specifying the acceptable error. We now have afunction to decide which output matches the specific input. The difference e

    and the network output y, sends out an error-signal, this is used to adjustthe free parameters, to make the difference disappear statistically.

    The image above shows a two graphs, one is the f(x) function, the other theapproximated F(x) function.

    F(x)-f(x)

  • 8/13/2019 Soft Computing PDF

    95/103

    System identification

    Unknownsystem

    NN

    Inputvector Xi

    di

    Yi

    ei

    I

  • 8/13/2019 Soft Computing PDF

    96/103

    Inverse system

    f(.)di

    Yi

    ei

    Inversemodel

    Inputvector Xi

    X=f -(d)

    System output

    Model output

    Control J= yk/ uj

  • 8/13/2019 Soft Computing PDF

    97/103

    Yei

    Controller

    Reference input

    System outputPlant input

    Plantdi

    Unity feed back

    Indirect learning: Using actual input output measurement on the plant Aneural network model is constructed to produce copy of it. This model isinturn used to provide an estimate of the Jacobian matrix J. The partialDerivatives constituting this jacobian matrix are subsequently used in errorcorrection learning algorithm for computing the adjustments to the freeparameters of the NN

    j

    Direct learning: The sign of the partial derivatives are generally known andusually remain constant over the dynamic range of the plant. This suggest thatwe may approximate these partial derivative by their individual signs. Theirabsolute values are given a distributed representation in the free parameter ofthe neural controller. The neural controller their by enabled to learn theadjustments to its free parameters directly from the plant

    Filtering

  • 8/13/2019 Soft Computing PDF

    98/103

    Filtering

    The term filter refers to a device or algorithm used to extractinformation about a prescribed quantity of interest from a set ofnoisy data. We as humans are also able to perform filtering, in manysituations we are forced to isolate one voice from hundreds ofothers, in order to get the information we need. Neural networkscan, just as the brain, be trained to remove noise from various

    sources with great results.

    Filter can be used to perform the following

  • 8/13/2019 Soft Computing PDF

    99/103

    p g

    information processing tasks

    Filtering: This task refers to the extraction of information atdiscrete time n using data measured up to and including time n

    Smoothing: The information about the quantity of interest need notbe available at time n and data measured later time n can be used in

    obtaining this information. Thus there is a delay in producing results.The results may be more accurate than filtering.

    Prediction This task is the forecasting side of informationprocessing. The aim here is derive information about what the

    quantity of interest will be like at some time n + n0in the future forsome no, by using data measured up to and including time n.

    Nonlinear prediction

  • 8/13/2019 Soft Computing PDF

    100/103

    Nonlinear prediction

    NN

    X(n-T)

    X(n-2T)

    X(n-mT)

    X(n)

    Estimated value ofx(n)

    Beam forming

  • 8/13/2019 Soft Computing PDF

    101/103

    Beam forming

    Beamforming is a spatial form of filtering and is used todistinguish between the spatial properties of a targetsignal and back ground noise. The devise used to do thebeam forming is called beamformer

    Beamforming is commonly used in radar and sonarsystems where the primary task is to detect and track atarget of interest in the combined presence of receivernoise and interfering signal( e. g. Jammers). This task iscomplicated by two factors

    That target signal originates from an unknown direction

    There is no a priori information available on the interferingsignals

    Generalized sidelobe canceller

  • 8/13/2019 Soft Computing PDF

    102/103

    Generalized sidelobe canceller

    w1

    NNSignalBlocking

    Matrix Ca

    w2

    wm

    Desired responsed(n)

    Out puty(n)

    x(n)

    u1(n)

    u2(n)

    um(n)

  • 8/13/2019 Soft Computing PDF

    103/103

    An array of antenna elements:

    A linear combiner: Acts as spatial filter characterized by a radiation pattern

    Signal blocking matrix:

    The function of which is to cancel interference that leaks through thesidelobes of the radiation pattern of the spatial filter representing linearcombiner.

    Neural network