Summary of a neural model of human image categorization
-
Upload
jin-hwa-kim -
Category
Technology
-
view
142 -
download
1
description
Transcript of Summary of a neural model of human image categorization
Summary of A Neural Model of
Human Image Categorization
Methodology of Cognitive Science
Jin Hwa KimCognitive Science ProgramSeoul National University
What We Will See
1. Computational Neural Model
- Leaky integrate-and-fire (LIF) neuron model- Deep autoencoder- Circular convolution
2. How classes of visual objects are represented in the brain?
- Prototype-based (Posner & Keele, 1968)- Exemplar-based (Regehr & Brooks, 1993)
Leaky integrate-and-fire (LIF) neuron model
- One of biological neron models (spiking neuron model)
LIF Neuron Model
[Gerstner and Kistler, 2002]
Deep Autoencoder
Deep Autoencoder
Principal Component Analysis
direction of first principal component i.e. direction of greatest variance
Deep Autoencoder
Specialized neural network
- Try to make the output be the same as the input in a network with a central bottleneck
input vector
output vector
code
encoding weights
decoding weights
semantic pointer
Deep Autoencoder
Solving optimization problem
- Use unsupervised layer-by-layer pre-training.- LIF instead of RBM
784 ! 1000 ! 500 ! 250 30 linear units 784 " 1000 " 500 " 250 We train a stack of 4 RBM�s and then �unroll� them. Then we fine-tune with gentle backprop.
W1 W2 W3
W1T W2
T W3T
4W
TW4
[Hinton & Salakhutdinov, 2006]
Circular Convolution
Store semantic pointers
- Holographic reduced representations using compositional distributed representation
Categorization process
circular convolution operator
[Plate, 2003]
Visual Categorization Model
Posner & Keele, 1968
Prototype theory
- It was designed to test whether human subjects are learning about class prototypes when they only ever see distorted examples.
Figure 3: Sample stimuli for Experiment 1, modelling a classic study by Posner & Keele (1968). The dot patterns are created by distorting three randomly drawn prototype images (left) with low (centre) and high (right) levels of noise. Subjects are trained to classify a set of twelve high-distortion patterns and tested without feedback on the same prototypes at different distortion levels.
Results
Posner & Keele, 1968
Figure 4: Comparison of human and model performance for Experiment 1. The model is able to account for human results when presented with the schema, low distortion (5), and high distortion (7) patterns. Occasional random errors by human subjects may explain the discrepancy on training examples. Error bars indicate 95% confidence intervals. Human data from Posner & Keele (1968).
Exemplar theory
- Analytic vs. Perceptual similarity
Regehr & Brooks, 1993
Figure 5: Sample stimuli for Experiment 2, modelling experiment 1C of Regehr and Brooks (1993). (Left) Images are composed of interchangeable (composite) feature manifestations. (Right) Images expressing the same attributes are drawn in a more coherent (individuated) style. Regehr & Brooks (1993) drew a distinction between good transfer and bad transfer test stimuli. A test stimulus is a good transfer case when the addition or removal of spots matches a training case with the same label, and a bad transfer case if adding or removing spots matches a training case with the opposite label. (Adapted from Regehr & Brooks (1993) Figure 2A).
Results
Figure 6: Comparison of human and model performance for Experiment 2. Our model accounts for the key difference in human performance on the good transfer (GT) versus bad transfer (BT) pairs for the individuated stimuli. Error bars indicate 95% confidence intervals. Human data from Regehr & Brooks (1993).
Regehr & Brooks, 1993
Biological vs. Artificial neuron model
Double-reduced representation
- Deep autoencoder- Circular convolution
Discussion