- ...Gasser
- To whom correspondence should be addressed
- ...BLOCK
-
We will use uppercase for concepts, italics for
linguistic forms, and double quotes for utterances.
- ...representations.
-
We have no reason to believe, however, that the conclusions we reach
will not generalize to other representational schemes.
An alternative, for example, is a variant of localized
encoding in which units on either side of the most highly activated unit are
also activated, in inverse proportion to their distance from the activated
unit.
A version of the present network using such a scheme trained on the data
generated for Experiment 3
below exhibited the same advantage for compact over elongated
categories as was found with thermometer encoding.
- ...here.
-
Increasing the number of units in the hidden layer of the
network both speeds up performance and leads to improvement in
the asymptotic level of performance.
- ...terms.
- As we will see in subsequent experiments,
the noun advantage in the network does not depend on there
being only two terms for each adjective dimension.
- ...(
-
For statistical tests here and in Experiments 2--5,
we treated each run of the network as a separate
subject.
- ...nouns.
-
An initial difference in
learning but ultimately equal and near perfect learning of both nouns
and adjectives is achieved with larger hidden layers.
- ...cases.
- All
of the mean activations are negative because for this experiment, the
network learns to strongly inhibit all but the right response for each
training instance,
and for the test patterns, there is no ``right'' response from among
the trained categories.
- ...(
-
For the analysis of variance, there were two factors, input linguistic
context (noun or adjective) and average activation over output
units by meta-category (noun or adjective).
There was only one ``subject'' (network run) in this experiment, but
there were 18 instances of each of the four combinations of the factors.
Michael Gasser
Fri Dec 6 13:15:34 EST 1996