KEMBAR78
Neural Networks Basics | PDF | Artificial Neural Network | Nervous System
100% found this document useful (1 vote)
160 views25 pages

Neural Networks Basics

This chapter introduces neural networks and their basic concepts. It defines a neural network as a massively parallel distributed processor made up of simple units that can store experiential knowledge through a learning process. Knowledge is represented by connection strengths between neurons called synaptic weights. The chapter outlines key benefits of neural networks and describes the basic structure and function of the human brain as a neural network, as well as comparisons between biological and artificial neural networks. It introduces common neuron and activation function models used in neural networks.

Uploaded by

Ranit Nath
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
160 views25 pages

Neural Networks Basics

This chapter introduces neural networks and their basic concepts. It defines a neural network as a massively parallel distributed processor made up of simple units that can store experiential knowledge through a learning process. Knowledge is represented by connection strengths between neurons called synaptic weights. The chapter outlines key benefits of neural networks and describes the basic structure and function of the human brain as a neural network, as well as comparisons between biological and artificial neural networks. It introduces common neuron and activation function models used in neural networks.

Uploaded by

Ranit Nath
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 25

Neural Networks 2nd Edition

Simon Haykin

Chap 1. Introduction

柯博昌
What is a Neural Network

 A neural network is a massively parallel distributed


processor made up of simple processing units, which
has a natural propensity for storing experiential
knowledge and making it available for use.
 Knowledge is acquired by the network from its
environment through a learning process. The
procedure performing learning process is called a
learning algorithm.
 Interneuron connection strengths, known as synaptic
weights, are used to store the acquired knowledge.

2
Benefits of Neural Networks

 The computing power of neural networks


– Massively parallel distributed structure
– Ability to learn and therefore generalize.
 Using neural networks offers following properties:
– Nonlinearity – Fault Tolerance
– Input-Output Mapping – VLSI Implementability
– Adaptively – Uniformity of Analysis and Design
– Evidential Response – Neurobiological Analogy
– Contextual Information
 Supervised Learning: Modifying the synaptic weights
by applying a set of training samples, which constitute
of input signals and corresponding desired responses.
3
Human Brain - Function Block

 Block diagram representation of human nervous


system Forward
Neural Response
Stimulus Receptors Effectors
Net.
(Brain)
Feedback
 Receptors: Convert stimulus from the human body or
the external environment into electrical impulses that
convey information to brain.
 Effectors: Convert electrical impulses generated by
brain into discernible responses as system outputs.
4
Comparisons: Neural Net. Vs. Brain

 Neuron: The structural constituents of the brain.


 Neurons are five to six orders of magnitude slower
than silicon logic gates. (e.g. Silicon chips: 10-9 s,
Neural Event: 10-3 s)
 10 billion neurons and 60 trillion synapses or
connections are in the human cortex.
 The energetic efficiency
– Brain: 10-16 joules per operation per second.
– Best Computer today: 10-6 joules per operation per second.

5
Synapses

 Synapses are elementary structural and functional


units that mediate the interactions between neurons.
 The most common kind of synapse is chemical
synapse.
 The operations of synapse:
– A pre-synaptic process liberates a transmitter substance
that diffuses across the synaptic junction between neurons.
– Acts on a post-synaptic process.
– Synapse converts a pre-synaptic electrical signal into a
chemical signal and then back into a post-synaptic electrical
signal. (Nonreciprocal two-port device)

6
Pyramidal Cell

7
Cytoarchitectural map of the cerebral cortex

8
Nonlinear model of a neuron
Bias
x1 bk
wk1
Activation
function
Input x2 wk2 Output
signals S vk j(×) yk
...

...

Summing
xm wkm junction
Synaptic
weights
m
uk   wkj x j vk  u k  bk yk  j (vk )  j (uk  bk )
j 1

Let bk=wk0 and x0=+1 vk   wkj x j and yk  j (vk )


m

9 j 0
Nonlinear model of a neuron (Cont.)

Induced bk>0 Fixed input


wk0=bk (bias)
x0=+1 wk0
local Bk=0
field, uk Bk<0 x1 wk1
Activation
function
Input x2 wk2 Output
signals S vk j(×) yk

...

...
Summing
Linear combiner’s xm wkm junction
output, uk Synaptic
weights
(including bias)

Affine transformation produced Another Nonlinear model of a neuron


by the presence of a bias

10
Types of Activation Function
1.2
Threshold Function 1 j (v )
0.8
1 if v  0
j (v )  
0.6
0.4
0 if v  0 0.2
0
-2 -1.5 -1 -0.5 0 0.5 1 1.5 2
v
Piecewise-Linear Function 1.2
j (v )
v  1/ 2
1
1
v  1 / 2  v  1 / 2 0.8
j (v )   0.6
0.4
0 v  1 / 2 0.2
0
-2 -1.5 -1 -0.5 0 0.5 1 1.5 2
v
Sigmoid Function 1.2
j (v )
j (v ) 
1 1

1  exp( av )
0.8
0.6 Increasing
0.4 a
a is the slope parameter 0.2
11 0
-8 -6 -4 -2 0 2 4 6 8
v
Types of Activation Function (Cont.)

 The activation functions defined above range from 0


to +1.
 Sometimes, the activation function ranges from -1 to
+1. (How to do?)
Assume the activation function ranging from 0 to +1 is denoted as j(×),
ranging from -1 to +1 is denoted as j’(×)
 j’(×)=j(×)*2-1
Notes: if j(v)=sigmoid function

j (v ) 
1
* 2 1
1  exp( av )
1  exp( av )
  tanh(v)
12 1  exp( av )
Stochastic Model of a Neuron

 The above model is deterministic in that its input-


output behavior is precisely defined.
 Some applications of neural network base the
analysis on a stochastic neuronal model.
Let x denote the state of the neuron, and P(v) denote the probability of
firing, where v is the induced local field of the neuron.
 1 with probability P(v)
x
 1 with probability 1 - P(v)
A standard choice for P(v) is the sigmoid-shaped function. T is a
pseudo-temperature that is used to control the noise level and
therefore the uncertainty in firing.
P(v ) 
1
13 1  exp( v / T )
Neural Network  Directed Graph

wkj
Synaptic Links xj yk=wkjxj

j(×)
Activation Links xj yk=j(xj)

yi
Synaptic Convergence yk=yi+yj
(fan-in)
yj
xj
Synaptic Divergence
xj
(fan-out)
xj

14
Signal-flow Graph of a Neuron

x0=+1

x1 wk0=bk
wk1
wk2
vk j(×)
x2 yk
...
...

wkm

xm
15
Feedback

 Feedback plays a major role in recurrent network.

xj’(n) A
xj(n) yk=j(xj)

yk(n)=A[xj’(n)] xj’(n)=xj(n)+B[yk(n)] where A and B act as operators

yk (n ) 
A
1  AB
 
x j (n ) A/(1-AB) is referred as closed-loop operator,
AB as open-loop operator.

In general, ABBA
16
Feedback (Cont.)

Let A be a fixed weight, w; and B is a unit-delay operator, z-1


A

w
1  AB 1  wz 1
 w 1  wz 1
( )1

( )  w z

1 1 l Use Taylor’s Expansion or Binomial
 1  wz l

l 0
Expansion to prove it.

A
  w wl z l
1  AB l 0

   

yk (n )  w wl z l x j (n )  z  l x j (n )  x j (n  l )
l 0

 

yk (n )   wl 1 x j (n  l )
l 0

17
Time Responses for different weight, w

yk(n) yk(n) yk(n)

wxj(0) w=1 w>1


w<1

wxj(0) wxj(0)

n n n
0 1 2 3 4 5 0 1 2 3 4 5 0 1 2 3 4 5

Conclusions:
1. |w|<1, yk(n) is exponentially convergent. System is stable.
2. |w|1, yk(n) is divergent. System is unstable.

Think about:
1. What does the time response change, If -1<w<0?
2. What does the time response change, If w-1?
18
Network Architectures
MultiLayer
Single-Layer
Feedforward
Feedforward
Networks
Networks

Fully Connected: Every node


in each layer is connected to
every other node in the
adjacent forward layer.
Input layer
Output layer
Otherwise, it’s Partially
of source Connected.
of neurons
nodes

Input layer Layer of Layer of


of source hidden output
19 nodes neurons neurons
Network Architectures (Cont.)
Recurrent Networks
Recurrent Networks
with no self-
with hidden neurons
feedback loops and
no hidden neurons

z-1
Outputs
Unit-delay z-1
z-1 z-1 z-1 z-1
operators
z-1

z-1
Unit-delay
operators

Inputs

20
Knowledge Representation

 Primary characteristics of knowledge representation


– What information is actually made explicit
– How the information is physically encoded for subsequent
use
 Knowledge is goal directed.
 A good solution depends on a good representation
of knowledge.
 A set of input-output pairs, with each pair consisting
of an input signal and the corresponding desired
response, is referred to as a set of training data or
training sample.
21
Rules for Knowledge Representation

Rule 1: Similar inputs from similar classes should usually produce similar
representations inside the network.
Similarity Measuring:
(1) Using Euclidian distance, d(xi, xj) (2) Using Inner Product, (xi, xj)
Let xi=[xi1, xi2, …, xim]T Let xi=[xi1, xi2, …, xim]T
(x , x )  x
m
x j   xik x jk
1
 2
T
d (xi , x j )  xi  x j   (xik  x jk ) 
m 2
i j i
 k 1  k 1

1 xi
Similarity 
d (xi , x j )

cos( ) 
(x , x ) i j

xj
xi × x j xiTxj

22 If ||xi||=1 and ||xj||=1 d2(xi, xj)=(xi-xj)T(xi-xj)=2-2(xiTxj)


Rules for Knowledge Representation
(Cont.)

Rule 2: Items to be categorized as separate classes should be


given widely different representations in the network.
(This is the exact opposite of Rule 1.)
Rule 3: If a particular feature is important, then there should be a
large number of neurons involved in the representation of
that item.
Rule 4: Prior information and invariance should be built into the
design of a neural network, thereby simplifying the network
design by not having to learn them.

23
How to Build Prior Information into
Neural Network Design

 Restricting the network architecture though the use of local


connections knows as receptive fields.
 Constraining the choice of synaptic weights through the use of
weight-sharing.
6
v j   wi xi  j 1 ,
Ex:
j  1,2,3,4
i 1
Convolution Sum

Convolution Network

x1, …, x6 constitute the receptive


field for hidden neuron 1 and so
on for the other hidden neurons.
24
Artificial Intelligence (AI)

 Goal: Developing paradigms or algorithms that


require machines to perform cognitive tasks.
 AI system must be capable of doing:
– Store knowledge
– Apply knowledge stored to solve problems
– Acquire new knowledge through experience
 Key components
Representation
– Representation
– Reasoning
– Learning Learning

Reasoning
25

You might also like