Theses
Thesis/Dissertation Collections
1991
Algorithm animation and its application to artificial
neural network learning
Walter Bubie C.
Follow this and additional works at:
http://scholarworks.rit.edu/theses
This Thesis is brought to you for free and open access by the Thesis/Dissertation Collections at RIT Scholar Works. It has been accepted for inclusion
in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact
ritscholarworks@rit.edu.
Recommended Citation
September 18, 1991
Approved
by:
Algorithm Animation and its Application
to Artificial Neural Network Learning
by
Walter
C.
Bubie
A thesis, submitted to the Faculty of the Computer Science
Department,
in
partial fulfilhnent of the requirements for the
degree of Master of Science
in
Computer Science.
Professor Peter G. Anderson
Dr. James Wilson
Algorithm Animation and its Application
to Artificial Neural Network Learning
I, Walter
C.
Bubie, hereby grant permission to the Wallace Memorial Library, of
RIT, to reproduce my thesis
in
whole or
in
part. Any reproduction will not be for
corrunercial use or profit.
I
extendmy
sincerethanks to
Dr. James Wilson for
his
constantencouragementand
thought
provoking
suggestionsthat
helped
meldthe
different
disciplines
behind
this
thesis.
His
corrections of earlierdrafts
are alsovery
muchappreciated.
Many
thanks
to
Professor Peter
Anderson
for his
constantinterest in
the theme
ofthis work,
his
suggestionstowards
viewimprovements,
andhis
patience.Thanks
alsoto
Stan Caplan
ofthe
Design Resource Center /Human Factors
group
at
the
Eastman Kodak
Company,
for
his
willingnessto
give me morethan three
months off
from
workin
orderto
focus
oncompleting
this thesis.
Finally,
I
wishto thank
my
lovely
wife,
Doris,
whohas known
meonly
as aAlgorithm
animationis
a means ofexploring
the
dynamic
behavior
of algorithmsusing
computer-generated graphicsto
represent algorithmdata
andoperations.Research
in
this
field has focused
onthe
architectureofflexible
environmentsfor
exploring
small,
complexalgorithmsfor
data
structuremanipulation.This
thesis
describes
a projectexamining
two
relatively
unexplored aspects ofalgorithm animation:issues
of viewdesign
effectivenessandits
applicationto
adifferent
type
ofalgorithm,
namely back-propagation
artificial neural networklearning.
The
work entaileddeveloping
aframework
for
profiling
viewsaccording
to
attributes such as
symmetry, regularity, complexity,
etc.This framework
wasbased
on current researchin
graphicaldata
analysis and perception and servedas ameansof
informally
evaluating
the
effectiveness of certaindesign
attributes.Three
animated views weredeveloped
withinthe
framework,
together
with aprototype algorithm animation system
to
"run" each view and providethe
user/viewer
interactive
control ofboth
the
learning
process andthe
animation.Three
simple artificial neural network classifiers were studiedthrough
ninestructured
investigations. These investigations
explored variousissues
raised atthe
project outset.Findings from
these
investigations indicate
that
animatedviews can
portray
algorithmbehaviors
such asconvergence,
feature
extraction,
and
oscillatory
behavior
atthe
onset oflearning. The
prototype algorithmanimation system
design
satisfiedthe
initial
requirements ofextensibility
andend-user run-time control.
The
degree
to
which a viewis informative
wasfound
to
depend
onthe
combined viewdesign
andthe
algorithm variables portrayed.Strengths
and weaknesses ofthe
viewdesign
framework
wereidentified.
Suggested
improvements
to the
design
framework,
viewdesigns
and algorithmsystem architectureare
described
in the
context offuture
work.CR Categories
andSubject
Descriptors:1.1.2
[Algebraic
Manipulation]:Algorithms-v4naZj/sis of
Algorithms;
1.2.6 [Artificial
Intelligence]: Learning-Cownecfz'om'smandNeural
Nets;
1.5.2
[Pattern
Recognition]: Design Methodology-C/ass//zerDesign
andEvaluation;
1.3.6
[Computer Graphics]:
Methodology
andTechniques;
1.6.8 [Simulation
andModeling]:Types
ofSimulation-Animation;
Additional Keywords:
Algorithm
Animation,
Graphical Data
Analysis,
Interactive Program
Visualization.Acknowledgements
Abstract
1.
Introduction
1.1
The Problem
1
1.2
Thesis
Outline
3
1.3
Restrictions
andLimitations
3
2.
Background
2.1
Visualization
5
2.1.1
Program
andData Visualization
5
2.1.2
Algorithm Animation
6
2.2
Artificial
Neural Network Back-propagation
Learning
9
2.2.1
A
Brief
Overview
ofArtificial Neural
Networks
9
2.2.2
Multilayer
Networks
andBack-propagation
Learning
13
The Generalized
Delta
Rule
13
Strengths
andLimitations
of
Back-propagation
16
3.
Related Work
3.1
Visualization
andGood
Design
18
3.1.1
Gestalt
Psychology
19
3.1.2
Elements
of anEngaging
Image
21
3.1.3
Graphical
Data Analysis
21
Multivariate Data
Display
23
3.2
Algorithm Animation Systems
29
3.2.1
Algorithm Movies
29
3.2.2
Movie/Stills
30
3.2.3
Alladin
32
3.2.4
BALSA
33
3.2.5
Tango
35
3.2.6
Discussion
36
3.3
Artificial Neural Network
Systems
Providing
for Visualizations
39
3.3.1
Hinton
andKohonen
maps41
3.3.2
SunNet
44
3.3.3
P3
System
46
3.3.4
PDP
Software
47
3.3.5
NNanim
49
3.3.6
Back-Propagation
Dynamics
51
3.4
Empirical Work
onVisualizing
Algorithms
53
3.4.1
Static
Graphics
53
3.4.2
Dynamic Graphics
55
3.4.3
Discussion
56
4.
Project
Description
4.1
Motives, Objectives,
andChallenges
58
4.2
Project Overview
60
4.2.1
Approach
60
4.2.2
Neural Network
Classifiers
Explored
62
The XOR
Operation
62
Symmetry
Classifier
63
4.3
Project
Design
65
4.3.1
Animation View Design
65
4.3.2
Final
Views: Details
70
Draftsman
Plot
70
Texture
Map
View
73
Receptive
Field View
77
Energy
Plot
80
Discussion
81
4.3.3
Implementation Environment
83
5.
The
NetViz Prototype System
5.1
A Basic Tour:
Exploring
the
XOR Operation
85
5.2
System Details
90
5.2.1
Objectives
90
Algorithm Animation Specific Design Elements
90
Data Design
92
User Control
92
5.2.2
System Architecture
93
User Interface
93
Algorithm
Generator
99
Animation
View
Generator
100
Playback Controller
101
5.2.3
Implementation
Overview
102
5.2.4
External
Interfaces:
Control Files
104
Command File
104
Network Definition File
104
Pattern File
105
Animation
Definition
File
105
Weights
File
106
Record File
106
5.2.5
Major
Data
Structures
106
Network Structures
107
Color Table
108
Windows
andthe
Window List
109
View
"info"Structures
Ill
6.
Investigations
andObservations
6.1
Investigations
112
6.2
Observations
114
Investigation
1
114
Investigation
2
119
Investigations
122
Investigation 4
123
Investigations
124
Investigation 6
127
Investigation
7
129
Investigation 8
132
6.3
Observations
ofInvestigation
Variants
134
6.4
Observations
by
Other Viewers
135
6.5
Findings
Learning
Related
136
6.6
Findings
System
(View)
Related
138
View Effectiveness
138
Symbol
-VariableMapping
140
System
Interactivity
141
7.
Conclusions
andFuture Work
7.1
Conclusions
142
7.2
Suggested Future Work
146
7.2.1
Usability
Research
146
7.2.2
User
Interface
Design Research
147
7.2.3
System
Design Research
148
References
Glossary
Appendicies
A.
Generalized
Delta Rule /Back-Propagation
Algorithm
B.
NetViz User's Guide
C.
NetViz System Details
D.
Summary
ofInvestigations
(Table)
Introduction
1.1
The Problem
Algorithm
animationis
a visualizationtechnique
for
exploring
the
dynamic
behavior
ofalgorithmsusing
computer generated graphicsto
representalgorithm generated
data
and operations.Exploring
algorithmsthis
way
canpotentially
providethe
viewer withinsights into
the
behavior
ofthe
algorithmthat
are not possibleby
conventional means ofreading
valueslisted
atkey
junctures
ofthe
program.Algorithm
animationhas been
usedto
explore andteach
avariety
offundamental
algorithmsbased
onsorting
andsearching,
andhas
proven usefulin
algorithmoptimization,
debugging
and general research[Brow88a].
The
potentialbenefits
of algorithm animationhave
resultedin
the
development
of several general purpose prototype systems aimed atfacilitating
animated view construction
by
the
end-user[Hyrs87],
[Bent87],
[Stas90].
One
arearelatively
unexploredby
algorithmanimationis
artificial neuralnetworks
(ANN). ANN
learning
is
the
process ofsimulating
the
way
interconnected
brain
neuronslearn
to
recognize aspecific class ofpatterns,
generally using
simplified mathematical models ofbrain
structures.This is
a subject ofintense
interest
to
computerscientists, neurophysiologist,
and avariety
of other
disciplines,
asindicated
by
the
publishedliterature
describing
newalgorithms and applications.
This
researchis
providing
solutionsto
problemsthat
wereincompletely
solvedfor
the
last
twenty
years.Much
ofthe interest
is
also attributableto the
curiosity
ofhow
and whatANNs
learn [Tour89]. Unlike
expertsystems,
anANN
cannot provide atrace
ofits
reasoning.
The
nearestanalog
ofknowledge
in
anANN,
comparedto
an expertsystem,
is
encodedin
the
weights ofits
interconnections.
These
aresimply
realnumbers,
whichout of some context areusually
meaningless.Understanding
such
ANN internals
typically
providesthe
key
towards
the
development
ofunderstanding
appearsto
requirea combinationofpenetrating
mathematical analysisandtrial
anderror,
"brute
force" experimentation.ANN
training1can
be described
as"seli-programrning."To
explorethe
inner
structureofANNs,
many ANN
researchersuse someform
of staticvisualization,
usually
depicting
the
system's stateat selected stages.In
contrast,
a conventional programis
typically
developed
withthe
aid of aninteractive,
dynamic debugger:
atool
for
confirming,
in
a stepwisefashion,
that
the
programbehaves
asexpected.
Since
anANN
only becomes
useful whentrained
successfully,
it
seems reasonable
that
ananalogoustool
canbe
ofbenefit for ANN
development.
It is
proposedhere,
that
algorithmanimation,
by
providing
interactive
anddynamic
graphicalexplorationofANN
learning,
is
such atool.
A
key
issue
not yet addressedin the
publishedliteratureon algorithm animationare
the
elementsthat
makefor
an"effective
view,"
i.e.,
a viewthat
triggers
insight
aboutthe
algorithmin the
viewer's mind.According
to the
publishedliterature,
the
way
aroundthis
issue is
to
iterate
onthe
design
of an animatedview,
for
aparticular
audience,
until a"good
design"is found
according
to
informal
criteriaof
the
designer
[Brow85],
[Stas90].
Although
it
canbe
postulatedfrom
reportedexperienceswithsystems such as
BALSA
[Sedg84], [Brow88a],
Tango
[Stas90],
and
Animus
[Duis86]
that this
approach enhancesthe
viewer'sunderstanding,
designers
couldbenefit
-especially
if
they
arethe
end-users-from design
guidelinesthat
maximizethe
probability
of asuccessful view.Similarly,
there
is
a need
for
aframework
for
describing
basic
elements of algorithm animationviews
in
away
that
the
effectiveness ofdifferent
views ofthe
same algorithm canbe
evaluated.This
thesis
describes
workthat
couples algorithm animation ofback-propagation
learning
withelements of"good
form"and methods ofmultidimensional
data
display
for
viewdesign. The
contributionsofthe
work are:a
framework for
informally
determining
the
effectivenessof algorithm animation viewsfor
neural networklearning
anextensibleprototype system
for exploring back-propagation
andpotentially
otherlearning
algorithmsthrough
different dynamic
views.1 Althoughthenormalcomplementto"learn"
is"teach,""train"
The
first
part ofSection
2 is
a generaldiscussion
on visualization and morespecifically
onalgorithm animation.The
second partis
an overview ofANNs,
with anemphasisonlayered
networkarchitecturesthat
usethe
back-propagation
learning
algorithm.Section
3 is
a review of related work: principles of gooddesign
according
to
Gestalt
Psychology
and statisticaldata
analysis,
algorithm animationsystems,
andANN
systemsproviding
visualizationmethods.
Section
4
is
the
projectdescription:
the
objectivesand approach.In
Section
5
the
NetViz
systemarchitectureis
described,
a prototype algorithmanimation system
for
back-propagation,
createdfor
this
research.Section 6 is
adiscussion
ofobservationsandfindings
madeduring
a set of specificinvestigations.
Finally,
conclusionsand suggestionsfor
future
researcharediscussed
in
Section
7.
Following
Section
7
arethe
references,
aglossary,
and appendices.The
glossary
containsterms
specificto the topics
ofthis work; their
first
occurrencein
the text
is
generally
depicted
in italics.
Three
appendicesinclude
details
ofthe
back-propagation
algorithm,
details
ofthe
NetViz
architecture and aNetViz
user'sguide.
1.3
Restrictions
and
Limitations
It is
notthe
purpose ofthis
projectto
formally
validatethe
value,
orlack thereof,
of
using
dynamic
visualizationto
gaininsight
aboutANN learning.
Rather,
the
purposeis
to
establishan approachtowards
formal
experiments,
a systemfor
future
work,
andinitial
empiricalguidelinesfor
designing
good views.A
key
assumptionis
that
the
userofNetViz
has
a priorunderstanding
ofartificial neural
networks,
in
particular,
multi-layeredtopologies
that
usethe
back-propagation
learning
algorithm.Furthermore,
the
new useris
not expectedto
automatically be
ableto
interpret
visualizations,
but
through
enoughexperimentationwith
the system,
it is
expectedthat the
userwilldevelop
anappropriate
"visual
vocabulary" ofmeaningful
images.
The
NetViz
systemis
aprototype, tailored to
dynamically
visualizeANN
learning.
Although
designed
to
easily
annexother neural network paradigms aslearning
algorithm,
using
three
different
viewdesigns. The
prototypeis
designed
to
be easily
transported to
other graphics-basedplatforms;
however,
in
its
currentform
it is
dependent
onDigital Equipment Corporation's
(DEC)
kernel-based VAX Workstation
Software
(VWS).
Finally,
the
classificationANNs investigated in
this
projectare cases withknown
solutionsthat
are usedto
demonstrate
the
animatedviews ofthe
system.Finding
optimalarchitectures,
optimallearning
parameters,
etc.for
these
2.1
Visualization
2.1.1
Program
andData Visualization
Today's
graphics-capablecomputers(PCs
to
superminiworkstations)
areimportant
tools
for
visualizing large
volumes of multidimensionaldata. Two
terms,
program visualization anddata
visualization,
describe
arapidly advancing
domain
ofsoftwareandmethodsspecificto
computer-based visualization.Program
visualization(PV)
is
the
use of computer generated graphicsto
map
data
producedby
one or morelinked
algorithmsthat
describe
aknown
model withbuilt-in
bounds
[McCor87], [Neil89],
[Wrig90].
Data
visualization(DV),
generally employing
super/super-minicomputers,
is
the
application of computergraphics,
using
data
originating
from
sensorssetup
to
"watch" phenomenafor
which amodeling
algorithmis
unknown ordifficult
to
formulate.
DV
also encompasses visualizations ofdata
produced as a"by
product"
of an applied
algorithm,
asin
the
case offinite
element analysis of a modeled object.The
morefamiliar
examples ofPV
application arethose
createdusing
high
end/super mini
computers,
for
studying
turbulent
flow
andvortexformation
[Meir91],
collision of a star withablack hole
[Dunc90]
andvisualizing
chaotic networks[Pick88a].
Computer-aided
tomography (CAT)
volumerendering
[Fuch89],
[Schw90]
andmeteorology
[Hibb89]
are common examples ofdata
visualization.PV is
also pursued onless
powerful,
graphics-basedPC
/workstations,
involving
literature: PegaSys
[Mori85]
andSEE
[Baec86]
are systemsdepicting
codeusing
specialgraphicsand
typography
(static
graphics);
the
PV
Prototype
[Hero85]
andPECAN
[Reis85]
aresystemssupporting
codedevelopment
anddebugging
using
dynamic
codevisualization;
INCENSE
[Myer83]
uses static graphicsto
aidin
debugging
complexdata
structures;
andBALSA
[Brow88a],
Animus
[Duis86],
andKAESTLE
[Boec86]
aresystemsfor visualizing dynamic
data.
2.1.2
Algorithm Animation
More
research attention appearsto
be
directed
towards the
PV domain
ofdynamic
data
, morecommonly
referredto
asalgorithm animation(AA).
Algorithm
animationdisplays
aredynamic
viewsshowing
a program'sfundamental
operationsin
conjunction with views ofchanging
program variables.Operations
include
transformations
and accessesto
data
andto
alesser
extent,
flow-of-control.
In
mostAA
systems,
data
structures and program operations of"interest"
are mapped
to
graphical attributes such aslocation (in
two
dimensions),
line
weight,
symbolsize,
color,
etc.In
addition,
special statements are placedin
the
algorithmcode,
preceding
and/or
following
algorithmstatements
that trigger
changesto the
graphic attributesaccording
to
updatesin
the
data
structures ofinterest.
Thus,
during
the
normal execution ofthe
algorithm, the
rapidsequence ofchanging
graphics appear as an animation.Algorithm
animation was notoriginally
computer-based;
pioneering
examples areelaborately
producedfilms
[Baec81]
[Boot75].
However,
withcomputer-based
AA,
users arepotentially
activeviewers,
ableto
controlthe
temporal
flow
ofthe animation, the
mappings of graphic elementsto
data
structures,
andthe
assignments of graphicattributes.
The
predominantapplicationofAA
has
been in
the
instruction
ofintroductory
programming,
and algorithms anddata
structurescourses[Sedg84], [Gian86],
[Raml85].
It
has
alsobeen
appliedin
researchand analysis of algorithmsandprocesses,
asdescribed
in
[Lond85], [Duis86],
and[Brow88].
Brown
[Brow88a]
further
suggeststhe
use ofAA
generatedimages
as"technical
drawings"
of
data
Persistence
'synthetic
Transformation
incremental
y
history
direct
Figure 2-1. A
taxonomy
for
algorithm animationsystems.Brown
[Brow88]
defines
a cubictaxonomy
of prevalentdisplays
usedin AA
systems,
as abasis
for
automatic algorithm annotation.The
three
axesdescribe
the
attributes:content,
transformation,
and persistence(Figure
2-1).
At
oppositeends on
the
content axis aredirect
and synthetic.Direct
displays
areisomorphic
to
the
data
structuresdepicted;
a view of ahorizontal
series ofbars
atdifferent
heights mapping
the
contents of anarray
of valuesis
an example.Synthetic
displays
are abstractionsusually
of algorithmoperations,
or ofdata
that
is
a"by
product."
An
exampleis
ahistogram
showing
the
number oftimes
key
functions
of an algorithm were executed
(an
execution profile).The
transformation
axisdescribes
the
component of animationranging from
discrete
to
incremental displays. In
discrete displays
the
animationis
oftenperceived as
uneven,
reflecting
adirect
temporal
mapping
of algorithmdata.
Incremental
displays embody
smoothedtransitions,
eitherby
someform
ofartificial
interpolation
ofthe
visualizeddata,
orby
the
nature ofthe
algorithm.For
example,
Figure 2-2
as adiscrete
display
would animatethe
"exchange" eventof
sorting,
by
the
suddenswap
oftwo
highlighted
sticks.As
anincremental
display,
the
two
sticks would appearto
slide acrossthe
viewtowards
each othersAlgorithm:
Quick Sort
Interesting
Event: Exchange
Figure
2-2.An incremental
view of asorting
algorithm("sticks"
view)of
overlapping
updateframes
The
final
axis,
persistence,
rangesfrom
current andhistory. Displays
characterizedas current show
only
the
most recentdata
values andtherefore
projectthe
truer
form
of animation.Displays
that
arehistorical
maintainin
viewsomenumber of sequential update-frames.
More
specifically, in
a currentdisplay
each update-frame
is
rendered overthe
top
ofits
predecessor(thereby
erasing
it),
and
in
someforms
ofhistorical
views,
each update-frameis
rendered adjacentto
its
predecessor,
effectively
forming
atime-series
plot.Baecker
characterizesthe
domain
ofAA
asthe
combined use of"the
crafts ofgraphic
design,
typography,
animation andcinematography
to
enhancethe
presentation and
understanding
of computer programs..[Baec86:p325].
Asserting
that
"presentation is
enhanced" overthe
usualtextual
form,
is
similarto the
justification
for
the
use of graphic analysis of statisticaldata.
However,
the
assertionthat
understanding
is
enhancedis
not wellverified,
sincelittle
testing
has been done
and published results are stillpreliminary
[Gian86],
[Cros89].
Brown
describes rewarding
experiences withBALSA
systems used over six years as alecture
aid andlaboratory
tool.
For
example,
explorations ofdynamic
Huffman
trees
led
to
animproved
variationofthe
algorithm.Similarly,
a variation ofShellsort
was"discovered"
during
observations of concurrentunderstanding;
however,
it is
also possiblethat these
claims are confoundedby
powerful and usefulsystemfeatures,
overly
"tailored" animatedpresentations,
and
the
fact
that the
problemsexplored are of a similar class(typically
algorithmsfor data
structurestaught
in
anintroductory
Computer Science
course).Some
ofthe
generaldifficulties
ofAA (and
visualizationin
general)
are[Brow88]:
Choosing
aninformative
viewespecially
whenusing
an animated view as acommunicationvehicle(e.g.,
a classroom).The
conceptual view ofthe
algorithm as pictured
by
the
creatormay
not coincide withthe
picture(if
one evenexists)
in the
viewer'smind.Complicating
a view'simplementation is
the
needto
know
during
viewdesign,
the
behavior
and potential range of valuesfor
eachdata
variable.Capturing
operations whichis
the
basis
ofAA;
algorithm operationsdo
notnecessarily
correspondto
each access or modification ofthe
algorithm's
data
structures.Accessing
a particular variablehas
different
"meanings"
at
different locations
in
the
code.Real-time
performance can sufferconsiderably
from
simply monitoring
algorithm
data
structures,
andcomputing
synthetic viewdata.
2.2
Artificial Neural Network Back-propagation
Learning
2.2.1
A Brief
Overview
ofArtificial Neural
Networks
ANNs
are named afterthe
networks of nerve cellsin the
brain. Although
lacking
in
many
ofthe
operationaldetails
of natural neuralnetworks, the
computing
modelsthat
have been
developed
andthat
arebeing
researchedform
astimulating domain
of parallel computing.ANN
are good at patternmatching,
patterncompletion,
and pattern classification.They
also contributeto the
understanding
ofneurophysiologicalphenomena,
as reportedin
[Gros85],
[McCle86],
and[Zips86b]
ANN
systems are comprised ofthe
following
basic
elements[Rume86a]:
A
set ofprocessing
units(or
nodes)
Weighted
connectionsbetween
unitsA
propagation ruleA
state of activationAn
outputfunction for
each unitA
learning
ruleAn
operating
environmentThe
relationship among
these
elementsis illustrated in Figure
2-3.
Given
the
parallelism of networkprocesses(i.e.,
learning
andrecall),
it is
convenientto
describe
the
network state at a given pointin
time.
The
set of activation valuesin
each of
the
processing
unitsbest
representsa network state.Activation
values change overtime
according
to the
unit's outputfunction,
its
propagation,
andactivation rules.
Typically,
the
outputfunction
and activation rules areexpressed
in
someform
ofthresholding
function,
which endows units withexcitory,
inhibitory,
and neutral properties.Process UnitStaleDescription
-m 0 ^8 +m
Figure
2-3.An
artificial neuron.A
propagation ruledetermines
the
effect ofthe
network connectedto
a particularunit.
This
involves
the
patternof connectionsbetween
units,
which atany
time
constitutes what
the
systemhas
learned.
Learning
is
fundamental
in
neuralsystems; through the
process ofrepeatedly presenting input
patternsand,
in
certain
types
ofnetworks,
associatedtarget patterns,
the
connection weights areprocessing
unit;
positive weights contributeexcitory
effects.Weights
near zeroeffectively
"shut
off"
the
presenceofthe
connectionbetween
two
units.Over
time, therefore,
the
network patternbecomes
"specialized"to the
input
patternspresented
up
to
somemoment.In
away,
aform
ofknowledge
develops.
Many
learning
rules aredescendants
ofHebbian
learning,
modifiedto
overcomelimitations
causedby
non-orthogonalinput
patterns andthe
lack
of errorcorrectionmechanisms.
Hebbian
learning
is
defined
as[Rume86a:p36]:
Adjust
the
connection strengthbetween
unitsA
andB
in
proportion
to the
product oftheir
simultaneousactivation.The
variety
ofarchitecturesbased
on connectionrules,
propagationrules,
andlearning
rulesthat
have been
explored canbe
taxonomized in
several ways.One
way
is
shownin Figure 2-4.
This
taxonomy
is first
divided
between
two
general networktopologies:
fully
interconnected,
single"layer"
and
layered
networks.Each
topology
category
is further
divided
by
the type
oflearning
rule:unsupervised or supervised
learning. Networks
that
learn
unsupervised arepresented
training
input
patterns without associated"correct
answers"
(target
patterns).
Networks
that
learn
superviseddo
soby being
presentedthe target
pattern associated with each
input
pattern presented.In
the
category
offully
interconnected
topologies,
to
this
author'sknowledge,
only
unsupervisedlearning
modelshave been
explored.Furthermore,
whatlearning
affects,
differs
among
the
predominant networktypes
in
this
category:in
Hopfield
networks(and its
derivatives)
learning
resultsfrom
adjustmentsto
activation valuesin
processing
units;
in
aCarpenter/Grossberg
networkthe
connectionweights are adjusted.Networks
ofboth
learning
models existfor layered
topologies.
Only
ANN:
fully
interconnected
units/singlelayerlearning:
unsupervisedinput: binary:
input:continuous:
Hopfield
(1981),
Hopfield derivatives:Hamming, Boltzman,
Mean FieldTheory,
Carpenter/Grossberg
Hopfield
(1984)
learning:supervised
(?)
layersof units
learning:unsupervised input:
binary
input:continuous:(?)
Kohonen self-organizingnet;
counterpropagation;Neocognitron
(1980)
learning:supervised input: binary:
input:continuous:
simpleperceptron(two
layer,
linearthreshold)
multilayer network(nonlinear
threshold)
Figure
2-4. Ataxonomy
of artificial neural networks.Rumelhart
provides amore general classification of commonlearning
models[Rume86a:p54]:
associativelearning
andregularity
discovery. Associative
learning
networks
learn
to
produce aparticularactivation patternon one set of nodes wheneveranother,
associated patternis
presented on another set.In
general,
such networks
have
fully
connectedtopologies,
designed
to
store relationshipsbetween
two
patternsin
adistributed form
among
the
non-input units1.Regularity discovery
networkslearn
to
respondto
distinctive (and
possibly
subtle)
features
in the
input
patterns.Networks
ofthis
type
in
generalhave
input
patternsagainin
adistributed
form.
2.2.2
Multilayer Networks
andBack-propagation
Learning
The
Generalized
Delta Ride
A
learning
algorithm ofinterest
andthe
choicefor
this
projectis
the
back-error
propagation algorithm
(or
morecommonly,
back-propagation
)
[Rume86a]
[Werb74:
citedin
Dayh90].
The
associated networktopology
has
three
layers:
Processing
units are classified as visible(at
the
input
and outputlayers),
andhidden.
Unit
connectionslink
the
input
unitsto the
hidden
units andthe
hidden
units
to the
outputunits1as
illustrated in
Figure
2-5. Intra-layer
connections are notdefined.
The
modelis
atype
ofregularity
discovery
which canbe
trained to
computearbitrary
input/output functions.
l
Calc'd Output Value
0
Connectionweight
Node InputValue
OutputLayer
Hidden Layer
InputLayer
Figure 2-5. Common
threelayer ANN
representation.The back-propagation
algorithmis
atwo-phase
algorithm.The first
phase entailsthe
forward
propagation ofinput
patterns presented atthe
input
layer,
withactivation
levels
calculatedatsuccessivehidden layers.
Activation is
calculatedfor
every processing
unitby
applying
a sigmoid(also
referredto
assemi-linear)
thresholding
function
to
its
summedinputs. At
the
outputlayer,
activations of all output unitsconstitutethe
computed outputfor
the
presented pattern.The
second,
back-propagation
phasebegins
withthe
comparison ofthe
calculated output and
target
outputassociated withthe
input
pattern.The
difference,
orerror,
forms
the
basis
for
adjusting
the
weights,
starting
withthose
weights
between
the
hidden-output layers
andending
withthose
between
the
input
and(first)
hidden layer.
The
objective ofthe
back-propagation
algorithmis
to
adjustthe
system weightsin
away
that
minimizesthe
total
system errorbetween
calculated andtarget
output patterns.
From
a simplifiedperspective,
the
systemerror,
with respectto
each weight
in
the
system,
canbe
expressed as a paraboloidfunction;
by
adjusting
the
weightin
proportionto
the
negative ofthe
derivative
ofthe
error,
the
resultis
the
descent along
the
parabolic error gradient.Since
there
are asmany
error gradients asthere
are weightsin
the system, the
back-propagation
algorithm uses a "generalized" gradient
descent
ordelta
ruleto
computethe
amount
by
whichto
adjust all weightsin
the
system.An
essential element ofthe
generalizeddelta
ruleis
that
the
derivative
ofthe
activation
function
exists.The
requirementthen
is
anonlinear,
continuous activationfunction
whichis differentiable.
A
sigmoidfunction,
shownin
Figure
2-6(a)
providesthis
requirementandis
typically
usedin
the
back-propagation algorithm.Since
it
is
continuous,
it
can also produce undesired activationlevels
that
straddlethe
"on"and"off"
states.
However,
the
derivative
of
the
function,
shownin
Figure
2-6(b),
adjuststhose
weightsentering
the
processing
unitthe
most,
sothat
subsequent activations are either closerto
zero or one.The
readeris
referredto
Appendix A
for
amoreformal description
ofthe
1-exp(-Ew(ij)*o(j)+
bias(j)
1Output
Energy
Activation
function derivative
0.25
(amount
applied toadjustweight)Input Signal
Figure 2-6(a). Sigmoid
activationfunction,
(b). Activation function derivative.
Often
employedin
the
back-propagation
algorithmis
abias
term
for
eachhidden
layer
and outputlayer
unit.Biases
areeffectively
incoming
weightsfrom
a unitwith a constant activation value of one.
During
the
back-propagation
phase,
they
are adjustedthe
sameway
that
weights are.They
provide a constantterm
in
the
weighted sum of eachin-feeding
unit.Computationally,
eachbias essentially
translates
the
sigmoidfunction
ofthe
particularunit,
thereby
adjusting
the
threshold
effect(Figure
2-7).
The
globaleffect ofbiases is
usually
animprovement in
learning by
the
network.Activationfunction
1
Output
Energy
(Low/
negativebias)
(High /positive
bias)
Input SignalTrairiing
a networkbegins
by
assigning
smallrandom valuesto
weights andbiases
(e.g.,
0.5).
The
set oftraining
and associatedtarget
patterns arethen
repeatedly
presented, the
errorfor
the
current patterncomputed,
and system weights adjusted sothat
the
total
system erroris
reducedon each subsequentiteration.
This
reductionofthe
system erroris
described
as convergence.Training
is
stoppedwhenthe total
errorreaches an errorcriterion,
atarget
minimumvalue,
at which pointthe
systemcorrectly
classifies alltraining
patternsandusually
previously
unpresentedpatterns.The
generalizeddelta
rule processconceptually
involves
the
error pointtraversing
a rough errorsurface
in
weight-space.The
error pointis
conceptually
the
system error atthe
network'spresent state.In
certaincircumstancesthe
error pointmay
become
"stuck"in
a surface pocketthat
is
abovethe
errorcriterion;
instead the
network
has
reached alocal
minimum.Convergence
is
conceptually
the
reaching
of a global minimum.The
probability
ofbecoming
stuckin
local
minimadepends
onthe
learning
rate,
the
momentumconstant,
andthe initial
weights set priorto training1.
The
learning
rate
conceptually describes
the
inherent
energy
ofthe
errorpoint;
attoo
low
a valuethe
error point canbe
thwarted
from
reaching
a global minimumby
the
roughness ofthe surface,
andtoo
high
a value can causeit
to
over shoot a pathto
a global minimum.
The
momentum constantconceptually
imparts
aninertia
effectto
the
errorpoint,
controlling
its direction
andthereby
enabling
training
athigher
(sometimes
double)
learning
rates.Finally,
starting
networktraining
withsmall random weights ensures
the
probability
ofreaching
a global minimumbecause
ofsymmetry
breaking. A
successfully
trained
systemgenerally has
unequal
weights,
whichmay be
difficult
to
form
if initial
weights are equal.(Consider:
the
delta
weight computedfor
a particular output unitis
addedto
each of
the
weightsleading
into
that
outputunit.)
Strengths
andLimitations
of
Back-propagation
The
predominant strength ofback-propagation
learning
is its
relativegeneral-purposeness.
Networks
canbe
configuredto
learn
avariety
ofpattern-mapping
problems,
for
example, text-to-phoneme
conversionrules[Senj87],
The
utility
ofthis
ANN
paradigmis
enhanced(or
hampered)
by
its flexible
architecture,
i.e.,
choicesfor
number oflayers,
interconnections,
processing
rules,
and
learning
constants.Back-propagation
learning's
two
maindrawbacks
areits
time
to train
larger
networks(more
so with respectto
training
setsize),
andits
tendency
to
become
stuckin
local
rrunima.Usually,
complex recognition problems(i.e.,
aproblem withmany
decision
planes)
require sets with ahuge
numberoftraining
patternsin
orderto
learn
well enoughto
generalize,
allof whichmustbe
presentedhundreds
to thousands
oftimes.
Also
affecting
training
time
is
eachadditionalinput
andhidden
unit whichincreases
the
number of weightsby
the
numberofunits
in
eachlayer,
all of which needto
be
adjusted.The
use of adjustablebiases
is
one method ofmiriimizing
the
occurrence oflocal
minima.Other
solutionsthat
addressboth
drawbacks
arepreprocessing
oftraining
patterns,
subdividing
the
input
layer
to
creatediameter-limited
receptivefields [Rume86b:
p348]
This
chapteris
a review ofpublishedworkthat
influenced
this
project,
including:
research on effectiveviewdesign;
noteworthy
algorithmsystems;
andvisualizationsof
ANNs.
The
reviewfor
effective viewdesign
encompassesprinciplesof
"good
form,"
from Gestalt
Psychology
andfrom
the
graphicaldisplay
ofdata. The
algorithm animation systems reviewed are selectedaccording
to the
uniquecontributions each makesto the
field.
Finally,
the
review ofANN
visualization coversthe
predominant static viewsfound
in
the
literature
andsystemscapable oflimited
dynamic
algorithm variable visualization.To
this
author's
knowledge,
there
areonly two,
unpublished papersspecifically
investigating
the
dynamics
ofANN
learning
using dynamic
visualization.3.1
Visualization
and
Good Design
The
way
a problemis
representedstrongly
influences
the
way
it
is
understoodand
ultimately (if
ever)
solved.According
to
Simon
[Simo81:
citedin
Boec86],
solving
a problem meansrepresenting
it in
away
that
its
solutionis
evident.One
common practice usedto
structure a problem andto
conceptualize solutionsis
through
visualthinking
[McKim72]. Simple
visualflunking
involves
sketching
mental
images
of problem elementsthereby
creating
a manipulablediagram for
generating
solutions.Flowcharts,
statisticaldata
plots,
design
renderings,
andstory boards
arejust
afew
examples of visualthinking.
According
to
several major studies on graphicalinformation
processing
in
humans
[Kahn81], Qule81], [Wick88], [McCor87], [Legg89],
the
human
visualinformation
processors are optimizedfor
extracting large
amounts ofinformation,
quickly,
from
non-symbolicgraphics andimages.
The
underlying
reasoning
is
that
graphics canbe
assimilated and processedin
parallel whiletext
requires serial andlogical
processing
Qule81]. In
assimilating
large
amounts ofCertain image designs
are optimalfor
a giventask,
andtargeting
suchdisplays
increases
the
chancesofgaining
insight
from
the
information.
Successful interpretation
of animage
(especially by
the reader)
partly determines
its
effectiveness.Effectiveness
in
turn
depends
onthe
viewer's visualvocabulary
[Brow88], i.e.,
the
set ofimagery
that
anindividual
learns
through
experiencesin
their
environment[Kapl82],
[Berl71],
andthrough the
properties ofthe
information
being
displayed.
3.1.1
Gestalt
Psychology
There
are no simpleheuristics
for
creating
effective graphicalrepresentations.However,
abasis for
effectivenessseems viablethrough the
applicationofPragnanz
or"good
form."This is
oneofthe
principles ofGestalt
Psychology,
whose
tenant
is
that
the
whole of animage imparts
adifferent
experiencethan
its
parts.
Pragnanz
describes
the
tendency
for
humans
to
perceivesimple,
symmetrical,
and regularforms
whenonly
parts ofthe
forms
aredisplayed
(i.e.,
"ambiguous"
forms,
that
are overlappedby
otherforms).
Good
forms
tend to
be
attractive,
"easier
to see",
in
the
sensethat their
image
components are notreadily
noticed
(i.e.,
harder
to
decompose)
[Hubb40], [Levi81],
[Kapl87].
The
basis
ofPragnanz is
a set often
primary
principles of perceptual grouping:1
.Proximity
theperception ofgrouping
by
elementsthatare closetogether.2.
Similarity
theperceptionofgrouping
by
elementsthatare
like
each other.3.
Continuity
theperceptionofa singleentity
withinambiguousforms.
o
o
o
o
o
Common fate
the
perceivedgrouping
ofimage
elementsthat
appearto
movein
the
same
direction
and speed.5.
Closure
the
tendency
to
see closedimages
asunitary,wholes.Symmetry
/
regularity
theattractiveness ofwhenthe
left half
ofanimage is
mirroredin the
righthalf
(the
vertical mirror axisis
perceivedthe
strongest ofmany
possible axes).7.
Connectedness
the
tendency
toperceiveany
uniform connected regionas a single unit.Common
region thetendency
of an observertogroup
elementsthatarelocated
withinthe
same perceived region.Complexity
thedescription
of animage in
terms
ofthe
numberofdistinct form
attributes
(e.g.,
curvature, number of corners, etc.).1 0.
Information
contentdescribes
thesimplicity
of afigure
andits
"predictability;
simplefigures have
alow information
content and are more"predictable"(recognizable
withoutseeing
thedetails).
. . .
WW %
Connectedness
(7)
and common region(8)
aretwo
principlesrecently
postulated[Rock90].
Rock
points outthat
connectednessmay be
afundamental
property
of each ofthe
originalfour
principles.The last
two
properties ofinformation
Extending
the
theory
ofGestalt
Psychology
is
the
theoretical
framework
ofKaplan
[Kapl82:
citedin
Leve88],
regarding
informational
richness and structure.Images
which areinformationally
richtend
to
be
engaging.Informational
richnessis
characterizedin
partby
ahigh
degree
ofcomplexity
in
animage.
Structure
relatesto the
symmetry
andregularity
in
animage.
Images
that
arecomplex,
regular and symmetricare consideredto
be
engaging
in
away
that
enhances
comprehension;
atthe
sametime, exploring
aninformationally
rich sceneis
also anon-trivial
task
[Leve88].
The
preferencefor
complexity
(i.e.,
informational
richness),
regularity
andsymmetry
in
images
is
reportedin
studiesby
Hubbell
[Hubb40],
andHolynski
[Holy85],
[Holy88].
Hubbell
studiedthe
elements of goodform
in
the
modifications made
to
ambiguous geometric objectsby
aesthetically
naivetest
subjects.
Holynski
andLewis,
in
a series ofstudies,
examinedthe
relationshipsbetween
viewer preference and several visual principlesfor
aknowledge
base
of aesthetic criteria.In
both
cases, the
findings
that complexity,
ratherthan
simplicity,
makesfor
an attractiveimage
seemscontradictory
to the
principles ofGestalt. A
likely
explanation comesfrom
Hubbell's
findings:
the
complexity
in
preferredimages is
"simplicity
withdifferentiation."
That
is,
from
an atomisticview,
animage is
complexby
its
countablefeatures.
However,
from
a globalperspective,
aslong
asthe
features form
a perceptualgrouping,
they
then
contribute
to the
image's
differentiating,
attractiveness.Several
studieshave determined
that
images
(as
well astext)
meaningfulto the
viewer are
better
assimilated, remembered,
andlater
recognized[Cros89]
[Chas74].
Furthermore,
studiesby
Shneiderman [Shne82:
citedin
Cros89],
examining
the
comprehensionof ordered anddisordered
codefragments,
andRock
[Rock90],
examining
images,
showthat
patterns of"good
form" are rememberedmore
quickly
andbetter
than those
with"bad
form."
3.1.3
Graphical
Data
Analysis
structureof analytical andphysical
data. Insight
is
gainedby
looking
for
patternsandrelationships
in
the
data;
proving
ordisproving
predictedstructures;
andby
suggesting
correctiveaction when useddiagnostically.
Chambers
et. al[Cham83]
discuss
principlesthat
contributeto
aneffectivedisplay
ofquantitativeinformation,
key
onesbeing:
spatially
(x,
y)
codedinformation
is
most effectivesymbol
encoding
by
varying
in
shape and sizeis "next
best"simple patterns
(formed
by
the
data)
are perceived morequickly
than
complex ones
large
clustersofobjects aremoreeasily
seenthan smaller,
isolated
onessymmetry, in
particularbilateral
andcircular,
make adata
view moreengaging
The
key
is
to
usethese
principlesto
maximize a view's visualimpact.
Bertin
[Bert81]
andChambers
et. al offerthese
guidelines:Exclude
elements commonto
alldata
sincethe
common,
non-differentiating
element(s)
reducesthe
impact
ofthe
image.
Similarly,
the
impact
of animage's
elements matchestheir
importance
in
the
context ofthe
analysis.Use
separate representationsfor
size and signin
symbol coding.If
the
magnitude of
the
mapped valueis
important,
then
it
shouldbe
size-coded,
andits
sign non-size-coded.Alternately,
the
symbol size can representits
full
value(i.e.,
smallest size:negative;
largest
size:positive), together
with acomplementing
sign coding.Scales
varying
length,
size,
anddirection
are superiorfor
symbols.These
attributescause variable
visibility,
providing
the
strongest perceptual effect of valuedifferences.
However,
eventhe
optimally
designed
symbols
depict
atbest
approximations,
because
a viewer'sinterpretation
of
the
symbolis
relative.Therefore,
different
symbols should alwaysbe
distinguishable.
Symbol
componentmapping
can affectthe
overall"texture"
of a symbol plot.
2.1.1).
However,
most ofthe
publishedworkusing
PV
andDV
sofar
has
focused
primarily
ontwo types
ofdata forms: "real
objects" and"scalar-valued" or
"vector-valued"
functions
oftwo to
four dimensions [Beck91].
The
type
of realobjects
studied,
have,
by
their
nature,
aspatialstructure(e.g.,
a3D
molecular model).Functions do
not needto
have
spatialdomains, however,
most ofthe
functions
studied are givenspatial structures.The
graphicaldata
display
ofboth
data
forms have
characteristically
involved 3D
volumetricrendering,
withphotorealistic
shading;
none ofthe
more established graphicaldisplay
techniques
(and
generally,
moreanalytical)
seemto
be
used,
such as multivariate scatterplots,
distributions,
andtwo-way
classifications[Cham83].
Becker
andCleveland
in
[Beck91]
proposebroadening
currentPV
andDV
to
include
establishedtechniques
of graphicaldata display. Their
argumentis
that
there
aremany data forms
that
are not real objects orfunctions,
andthat
many
have
morethan
four dimensions.
Moreover,
3D
objects(especially
abstract,
solidshapes)
are considered moredifficult
to
explore andinterpret
[Cham83],
because
they
requirethe
viewerto
mentally "navigate
around"the
partialimage
in
view.Human
depth
perceptionis
less
reliable(quantitatively)
than
spatial perceptionwithin a
two-dimensional
plane.In
fact,
knowledge
about"3D
things"
comes
from studying
them
from different
angles.Multivariate Data
Display
The
analysis of multidimensional or multivariatedata
posesinteresting
graphicaldisplay
problems sincethe two-dimensions
of aplotting
surface areinadequate.
Several
techniques
have been
devised
to
display
multivariatedata
in
two
dimensions:
symbol
encoding
generalized
draftsman
displays
multi-window
/
two-way
classificationscolor
coding
Symbol
encoding
uses abasic
icon
having
variousadjustable componentsto
which each of
the
many
data
variables(or
sources) is
mapped.The icon's
symbols,
Kleiner-Hartigan
tree
symbols[Klei81:
citedin
Cham83],
andChernoff
face
symbols[Cher73].
In
certaincircumstances,
symbolscanbe
"plotted,"
i.e.,
two
dimensions
ofthe
data
set arespatially
encoded,
andthe
remaining
dimensions
are encodedin
the
multi-componentsymbol.Figure
3-1
illustrates
the
use of a starsymbolto
depict
12
variablesof cardata.
The length
of eachray
is
proportionalto
the
variable'svalue.HONDA CIVIC PLYM. CHAMP RENAULT LE CAR
^>
^>
VV SCIROCCD OATSUN 210 VW RABBIT 0.
CHEVETTE DODGE COLT FIAT STRADA
MERC. MARQUIS DODGE ST.REGIS L. VERSAILLES
Figure
3-1.Use
of a star symbolfor
multivariatedata (from
[Cham83:
p!61]).Generalized draftsman's displays
[Tuke81:
citedin
Cham83]
derive
their
namefrom
the
top-front-side
view set of3D
objectscommonly
usedby
draftsmen.
However,
withthe
"generalized"form,
any
pair-wisedisplay
of multipledimensions
is
possible.Typically
plottedin
eachofthe
viewsis
ascatter plot ofthe two
data dimensions defined
for
the
particular view.An
exampleis
shown5 6 7 8 2 3 4
SEPAL LENGTH SEPAL WIDTH
2 4 6 PETAL LENGTH
1 2 PETAL WIDTH
Figure
3-2.A
generalizeddraftsman
display
(from
[Cham83:
pl46]).Multi-window
/two-way
classifications[Beck91],
[Cham83]
are useful waysto
display
a multivariatedata
sethaving
two
categorical variablesthat
arecross-classified.
That
is,
eachlevel
of each variable appearsexactly
once with eachlevel
ofthe
other.In
the
example multi-windowdisplay
in
Figure
3-3,
weight and price(of
cars)
arethe
cross-classified variables.The
display
is
constructed suchthat
the
number of rows and columns of"windows" equalsthe
number ofdata levels defined for
the two
variables.The
two
variables are examined withintwo
otherdata
categories.Often
in
this
display,
a row of windows atthe
top
and a column of windows onthe
right ofthe
main window matrixdisplay
the
aggregate of
the
corresponding
data
in
the
windowsbelow
andleft. The
top,
right-most,
windowis
a"super
plot"
a or
o
u LU or
< 0_ LU or
CD
l - .
12 3 4 5
1978 REPAIR RECORD
PRICE
Figure
3-3.A
two-way
classification plot(from [Cham83:
p!46]).Color
Coding
Information
encoding
through
coloris
next most"powerful"
to
geometricencoding,
i.e.,
very
large
anddistinctive
scales canbe
constructed.At
the
sametime,
the
use of colormapping
requiresbeing
very
selective.Color
discrimination
is
constrainedby
the
early
limits
ofvisualmemory
ratherthan
the
capacity
to discriminate
locally
between
tints
[Tuft90]. Color is
also sensitiveto
interactive
contextualeffects such as simultaneous contrast.Color
acuity
canvary widely
withina small sample ofpeople.Confounding
the
perceptualissues is
the
frequent
problemof monitorto
monitorcolorconsistency
and colorprecision
[Marc82].
Used correctly though,
coloris
ausefulattributefor domain
mapping.It is
inherently
multidimensionalwhendescribed
by
either components:hue,
saturation,
and value(a
spatial-perceptualclassification)
orred, green,
andblue
(additive
classificationfor
videodisplays).
Devising
a colormap
that
models acontinuousrange can
be
problematic.Ware
[Ware88]
analyzesthe
problems ofimagery, X-radiographs,
and astrophysical maps.The
colorin
these
tables
is
described
by
acontinuousplane or sequenceofthe
visible color space.For
his
experiment,
Ware
partitionedthe
kinds
ofinformation
portrayedin
univariate colortables
into
two
categories:metric andform.
Metric
information
denotes
the
quantity
stored at each point onthe
surface(e.g.,
of a map).Form
information
denotes
the
shapeorstructureofthe
surface.Therefore,
metric colortables
reveal"fracture
lines,"through
abrupt colordensity
changes,
andform
colortables
revealdifferentiations
of surfacevariations,
such as cusps andgradients.Ware
conducted a series of experimentsto
ascertainthat
certaincolor sequences aremore pronethan
othersto
induce
contrast related errorsin
the
viewer.Two
forms
of contrast related errors were of concern: simultaneous contrast and chromatic contrast.Simultaneous
contrastis
ahuman
perception problemwhere one patch of coloris
perceptually
"shifted"by
another adjacentorsurrounding
color patch.For
example,
agray
patchis
perceivedasdarker
on a whitebackground
andrelatively
lighteron ablack background.
Chromatic
contrastis
the correlate,
involving
color patches.Ware
found
that
monotonically varying
sequencesalong
eitherthree
primary
colorchannels1tend
to
cause
the
greatest errors.Examples
were"gray
scales,"
"saturation
scales,"such as
gray-to-red,
and red-to-greensequences.User
testing
indicated
that
non-monotonic sequences are resistantto
contrast erroreffects,
such asthe
purehue
scale(the
colorspectrum).Hue
based
sequences werefound
to
workwell as metric univariatetables;
however,
they
failed
asform-type
tables.
Based
ontests,
atable
constructedby
varying
along
the
achromatic(gray
scale)
channel(which is
suitableas aform-type
table)