Encoding the Levels of Information in Pictures and ... - APA PsycNET

2 downloads 0 Views 2MB Size Report
IDS, No. 2, 169-190. Encoding the Levels of Information in Pictures and Words. Alinda Friedman and Lyle E. Bourne, Jr. University of Colorado. SUMMARY.
Journal of Experimental Psychology: General 1976, Vol. IDS, No. 2, 169-190

Encoding the Levels of Information in Pictures and Words Alinda Friedman and Lyle E. Bourne, Jr. University of Colorado

SUMMARY From a levels-of-processing framework (e.g., Craik & Lockhart, 1972), we derive the idea that the levels of information implicit in a stimulus, such as its physical configuration, its name, and the category to which it belongs, may become available for subsequent processing at different times after stimulus onset. In particular, tasks which allow the use of physical codes should be performed more rapidly than those which require "deeper" information. There are two important implications here: The first is that the ability to use a code does not mean that that code has been "matched to" a representation in memory. The second is that "depth" effects (i.e., physical < name < conceptual) should be demonstrable within both pictorial and verbal materials. The approach as a whole may be contrasted with current dualcoding approaches (e.g., Paivio, Note 1), which, while they allow for different levels of meaning in both verbal and imaginal symbolic systems, seem forced to assume that certain types of conceptual information are more easily accommodated within the verbal system. This leads to the prediction that some kinds of conceptual information will not be directly available from pictures, but must instead become available to the subject via an interaction between the symbolic systems. We tested these assumptions with a speeded-inference task (Van Rijn, 1973), which has properties that allow for unambiguous interpretation of reaction time differences across stimulus materials which require processing to the same explicit depth. Briefly, we found that pictures yield faster inferences than do words when the same semantic information is required for performance (Experiments 1 and 2), and that physical information is available sooner than conceptual information for both pictures (Experiment 4) and words (Experiment 5). Moreover, some types of pictorial materials (e.g., representations of proper nouns) function symbolically to the extent that they do not have an advantage in discriminability over words (Experiments 6 and 7), unless physical features are added to them which are redundant with the conceptual information the subject needs to perform (Experiment 8). The results are best interpreted within a levels-of-processing framework, in which multiple codes or representations do not exist to be activated by the appropriate stimuli, but rather the stimuli themselves embody levels of information which are encoded and used as needed.

169

170

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

Hypotheses about the nature of knowledge often reduce to questions of representation; for example, In what form are experiences or skills stored such that they can be used for cataloging subsequent events ? From this perspective, it follows that "new" events are recognized by matching them to stored events, so that hypotheses about stimulus or event encoding are also reducible to questions of representation. A more recent and somewhat different analysis of encoding derives from a levels-of-processing framework (Craik & Lockhart, 1972; Posner, 1969). According to this approach, any stimulus (or experience) allows encoding along a multiplicity of informational dimensions, ranging from purely physical (e.g., lines, angles) to strictly cognitive qualities implicit in the stimulus by virtue of its relation to a concept or conceptual category. In this view, the "stimulus-as-coded" is a flexible entity which depends on the operations that a subject is required to and/or does perform with the abstracted information. It is assumed, moreover, that both the subject and the experimenter may "tap into" the abstraction process at different levels. Thus, the emphasis is shifted to questions of a more pragmatic nature; for example, To what depth does a stimulus have to be analyzed in order to be meaningful or usable in the given context ? In the present paper, we are concerned with the levels of information that are en-

This research was supported by National Institute of Mental Health Research Grant MH-14314 and by Research Scientist Award 1-K5-MH37497, and National Science Foundation Research Grant GB-340-77X. This research is Publication No. 57 of the Institute for the Study of Intellectual Behavior. We are extremely grateful to Vicki Hansen, Barbara Leonard, and Jeremiah Satterfield for running the experiments and for their helpful comments and insights throughout the research, and to Wendy Berning, Ronald Kellog, and Michael Yagemann for serving as subjects in Experiments 7 and 8. Thanks are also due to F. Craik for his patience with the first version of this manuscript. Requests for reprints should be sent to Alinda Friedman, Institute for the Study of Intellectual Behavior, University of Colorado, Boulder, Colorado 80302.

codable from pictures and words. We propose that a levels-of-processing analysis of encoding implies that (a) other things being equal, (e.g., task requirements, discriminability, etc.), deeper information should be equally derivable from pictures and words; but (b) when picture-word differences do occur, they are encoding phenomena which result because task requirements generally favor more discriminable or identifiable stimuli. The concept of multiple codes existing as stored representations against which current input is matched must be distinguished from the concept that levels of information are implicit in the stimulus event by virtue of the fact that the stimulus is both a physical entity (and thus will have physical features and relations among those features) and an instance of a concept or conceptual category. The former approach is best exemplified in the work of Posner (1969, 1973) and others (Posner, Boies, Eichelman, & Taylor, 1969). In the traditional view, most or all of the dimensions along which a stimulus may be coded already exist in stored representations. The processing of all the informational dimensions begins at stimulus onset, and recogtion or matching judgments rely on a "horse race" among the codes,, with physical codes generally finishing sooner (i.e., physical codes "find" their representations sooner than do name or conceptual codes). A subtly different levels-of-processing viewpoint is held by Craik and Lockhart (1972). They maintain that stimuli are processed successively (rather than in parallel) through physical, name, and conceptual stages and that memory representations are by-products of this encoding process. It makes little difference to the present argument whether one agrees with the notion that the codes are processed in parallel or whether, in fact, a serial model is correct; in either case, the order in which the codes become available is the same. The distinction between Posner and Craik and Lockhart which we wish to highlight is that within the latter approach, recognition does not logically entail a match with a memory representation. Instead, as the stimulus is processed through several levels, different types of in-

ENCODING PICTURES AND WORDS

formation become available for use in subsequent processing. The task a subject is faced with may or may not allow him to make use of a particular type of information. The idea that different levels of information are implicit in the stimulus is derived from the latter framework. We believe that, in fact, the processing of all available stimulus information begins at stimulus onset; however, we do not believe that recognition entails a "memory match" or, for that matter, that recognition is an appropriate construct for many of the tasks in which it is assumed to occur. Tasks which allow the use of physical information will generally be performed faster than those which do not, regardless of whether stimuli are pictures, words, or three-dimensional objects, simply because the physical information is available as a code sooner than name or conceptual information. However, the latter information is implicit in the stimulus and its availability does not necessarily rely on having named the stimulus or having processed it as a bundle of physical features. At this point, a digression is in order so that we may clarify what we mean by statements like "pictures may be identified sooner than words on the basis of physical discriminability." We believe that a stimulus is "identified" at a highly abstract level of meaning prior to, or possibly simultaneously with, the extraction of other information from it. Identification does not imply recognition, nor does it imply that such tacit knowledge is consciously accessible to the subject (see Turvey, 1974). A good analogy would be a computer programmer who is trying to get his program to execute a READ statement in FORTRAN which has mistakenly been formatted to expect real numbers ; instead, it finds alphanumeric information and aborts. What is puzzling to the programmer is that the compiler seems to "know" some very general and logically more abstract information about the data class prior to being able to read and/or recognize the specific data input. For the present purposes, by identification of the stimulus, we simply mean that the subject knows, in the sense that the compiler does, which stimuli have been presented and may

171

now proceed to encode those stimuli along physical, name, and conceptual dimensions. While this viewpoint is, of course, different from the Craik and Lockhart (1972) proposal that subjects encode successive levels of information from the outset, we believe that the theories are compatible to the extent that they both assume, as far as availability is concerned, that physical information will generally be available to consciousness sooner than name or conceptual information. Moreover, we agree with their assumption that memory representations of these different codes exist as by-products of the encoding process and are not necessary to the extraction of information from the stimulus. Both our own views and the approach outlined by Posner (1969) and Craik and Lockhart (1972) are in turn different from a dual-coding approach. While Paivio (1971, Chapter 3) has explicitly allowed for multiple levels of meaning in both imaginal and verbal systems (specifically representational, referential, and associative meaning), he has also stressed the functional differences between the systems. For example, The relations among the functions vis-a-vis the two symbolic systems can be informally summarized as follows: imagery is relatively better than the verbal system for representing and coping with the concrete aspects of a situation, with transformations, and with parallel processing in the spatial sense. The verbal system is superior in abstract and sequential processing tasks, (p. 38)

Also, Thus imagery is assumed to be specialized for the symbolic representation of concrete situations and events, speed and flexibility of transformational thinking. . . and parallel processing in the visual—spatial sense. The verbal system, on the other hand, is presumably characterized by its capacity to deal with abstract problems, concepts, and relationships, and for processing sequential information, (p. 434; italics added)

Moreover, in order to save the image from the classic criticism that images can neither mediate nor represent abstract, generic knowledge, Paivio (Note 1) resorts to the argument that general ideas may be represented via specific exemplars which are in some sense prototypical of a class. Indeed, Rosch (1975) has shown that priming with a category name facilitates physical identity

172

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

matches for good (i.e., prototypical) cate- ENCODING SEMANTIC INFORMATION FROM PICTURES AND WORDS gory members only. Importantly, she also found that while "basic objects" (e.g., dog, One mechanism that can conceivably proshoe, car) may indeed be representable by an duce performance differences across different "averaged shape" of all their particular in- levels of information (e.g., shape versus stances, this is not the case at the level of name versus category) is simply that deeper superordinate (generic) objects (e.g., ani- information is more difficult to encode or remal, clothing, vehicle) (Rosch, Mervis, quires different and more complex proceGray, Johnson, & Boyes-Braem, in press). dures than does physical information. This In other words, we may image a generic relationship should hold for pictures as well dog, which combines the essential visual as words; in particular, pictures should not properties and shape of all dogs; however, need to be "receded" in order to derive we do not have a comparable image of deeper information from them. In order to animal. demonstrate this unambiguously, we need a What is important for the present argu- task in which we can manipulate stimulus ment is that if a dual-coding approach is to materials along dimensions which we assume be viable, it must assume that superordinate require more or less analysis, in which the generic information about instances of con- information required for performance is excepts is probably mediated via the interaction plicitly denned and constant across stimulus between the imaginal and verbal symbolic materials, and in which there is some objecsystems; an instance or picture of a dog tive way of determining that the stimulus "matches" its stored imaginal representation, manipulations affect encoding and not any which may in turn "contact" the correspond- other operations required for performance. ing representation in the verbal system, from The same-different task used by Posner which we may derive information such as its (1969) and others to study processing time name, that it is an animal, that animals for various levels of information is deficient breathe, and so on. The implication is that in the latter two requirements. First, in certain types of conceptual information this task, the basis of the subject's matchshould not be directly available from pictures no-match decision is not explicit and must of specific instances and should at the very be inferred from reaction time differences least require commerce between the two sys- across stimulus materials. Second, there is tems. no sure way of separating stimulus encoding In order to test these proposals, we first from other possible mental processes inneed a task in which stimulus encoding is volved in the task, and we need a task in clearly separable from other operations. We which the stimulus-as-coded and the operathen need to demonstrate levels-of-processing tions which use that code are separable. effects within both picture and word stimuli. Figure 1 shows a model for a task which Accordingly, this paper is divided into two we believe satisfies these requirements—an sections: In the first, we present three ex- adapted version of the two-stimulus, subperiments which validate a model for a task ject-paced inference task (Daehler, 1972; that allows us to separate stimulus encoding Huttenlocher, 1962). In this task, the basis from subsequent operations and in which a of the subject's response is explicit in the repicture-word difference is shown to occur sponse itself, and there is a built-in check in the stimulus encoding stage. In the for the independence of stimulus encoding. second section, we present five experiments To illustrate the task, suppose there are in which discriminability is constant among four stimuli which vary on color (red or either picture or word stimuli and perform- blue) and form (circle or square) dimenance differences are found which are clearly sions. We present a subject with two attributable to different levels of processing successive stimuli which have a common required before task-relevant information is value on one dimension and a different available. value on the other (e.g., a red square and

ENCODING PICTURES AND WORDS

a red circle) and ask him to make an inference about which one of the four possible stimulus values is relevant. The cor-, rect value will depend on the predetermined assignment of each stimulus. For example, if we have told the subject that both stimuli contain the relevant value and are, therefore, positive instances ( + + ), then a red square followed by a red circle require the response red. However, the same stimulus sequence requires the response blue if the subject has been told that neither stimulus contains the relevant value ( ). Similarly, we may have the subject make -1— (square) and —I- (circle) inferences. Van Rijn (1973) required his subjects to make these inferences as quickly as they could, and showed that when type of inference is a between-subjects variable, .the order of difficulty (reaction time to solve correctly) is ++ < < - + = + -. The model in Figure 1 is based on these data. It assumes that once the relevant characteristics (of geometric designs) have been abstracted from the stimuli, ++ subjects respond with the •value common to both stimuli and subjects transform the common value to its complement and respond with the complement. We hypothesize that the only difference between ++ and conditions in Van Rijn's study is the additional mental operation that subjects must perform in the problem solving stage. All of the inferences or problem solving operations take place after the stimulus is encoded. We assume that the encoding stage involves the identification (as previously defined) of which two of the four possible stimuli have been shown and the subsequent abstraction of the characteristics

necessary for solution. Note that while we are here assuming that the task may be divided into discrete stages (i.e., encoding and problem solving), the encoding of the stimuli cannot. We simply assume that at some point after presentation of both stimuli, the subject has the information he needs in order to proceed. In the speeded-inference task, stimulus instances that are geometric designs require only superficial processing, since the solution to any problem is physically explicit in the stimuli. We wanted to design formally identical problems which would require a deeper level of analysis before solution-relevant characteristics would be available. To illustrate, suppose hippo, mouse, bus, and car are the instances, with attributes (LARGE or SMALL) and categories (ANIMAL or VEHICLE) as dimensions. The relevant values are characteristic of the stimuli but are not (necessarily) inherent in the stimulus configurations. The subject cannot determine the solution from any physical feature of either a picture or a word and must perform a deeper analysis to obtain the information he needs. Suppose we presented these "semantic" problems with the instances represented as either pictures or words. The first thing to notice is that the subject must abstract and explicitly name exactly the same information (i.e., one of the four values) regardless of stimulus modality. Thus, comparisons can be made across stimulus materials with the information required for performance kept constant. The simplest expectation based on the model of Figure 1 and an analysis of task requirements is that reaction time will be the same for pictures and words. If reProblem Solving

Stimulus Encoding

V IDENTIFY STIMULI

ABSTRACT RELEVANT CHARACTERISTICS

173

FIND COMMON VALUE

FIGURE 1. Process model for the speeded-inference task.

174

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

action time differences are observed, however, conclusions regarding encoding differences will be valid only if the difference in stimulus materials affects the encoding operations and nothing else. Our model for the task has a built-in check for whether the stimulus-as-coded can be separated from subsequent operations which use that code. Our argument is similar to Sternberg's (1969) demonstration that after practice in a memory search task, a visually degraded probe stimulus changes the y-intercept but not the slope of the reaction time function across different target set sizes. We propose that if changing stimulus materials (from pictures to words) affects stimulus encoding and not problem solving, the reaction time difference between ++ and inferences will remain constant across materials. This allows us unambiguously to assign differences between materials to stimulus encoding operations alone. On the other hand, if there is an interaction between stimulus materials and type of inference, then the stimulus-as-coded must somehow affect problem solving operations. If this is the case, we cannot unambiguously interpret a reaction time difference between pictures and words, and we would hesitate to argue that it arises because of differences in encoding operations per se. If the reaction time difference between + + and is constant, however, there are two hypothetical situations that can occur when subjects are required to abstract identical semantic information from either picture or word stimuli. The first possibility is that there will be no picture-word difference, and we would conclude that abstracting the same semantic information from either two pictures or two words requires similar if not identical processes for purposes of this speeded-inference task. This is the prediction of a literal levels-of-processing interpretation of stimulus encoding, since the task can only be completed after the stimuli are processed to the semantic level. The second possibility is that there will be a picture-word difference in reaction time. A "looser" levels-of-processing interpretation can account for the situation in which pictures are faster than words, using reasoning

analogous to that used for the "CC vs. Aa" effect in Posner and Mitchell's (1967) samedifferent reaction time task. The effect is commonly attributed to the more rapid availability of physical information than name information. In our task, since the semantic information needed for correct performance is fixed and explicit across materials, a picture-word difference in favor of pictures can result only from physical differences between pictures and words which cause pictures to be more easily discriminated from one another. Once the stimuli have been discriminated or identified, determination of the relevant characteristics from them should take a constant amount of time. On the other hand, a dual-coding model (Paivio, 1975, Note 1), with its emphasis on differences in the representations of pictures and words, can more easily accommodate a reaction time difference in favor of words. This approach currently assumes that the verbal symbolic system is better suited for representing generic information, such as the category to which a specific event (e.g., a hippo) may belong. Thus, if words are faster than pictures, it seems reasonable to conclude that pictures and words are equally discriininable, but that semantic information can be derived directly from words, while pictures have to be receded (i.e., the imaginal code must contact its verbal counterpart). Such a finding would also support a memory-match view of recognition. The experiments which follow were designed to try to distinguish among these alternatives. For purposes of clarity throughout the paper, we use the following terminology for the independent variables. Problem, or inference, refers to a particular inference condition ; except for Experiment 1, subjects made only ++ or inferences, and this was always a between-subjects variable. Stimulus type refers to all manipulations that result in stimuli which are different from one another in kind but may represent the same concept (e.g., pictures versus words, shapes versus names of states, words printed in color versus words printed in black ink, etc.). Within stimulus types, there are always only four unique instances (e.g., hippo, mouse, bus, and car). Two of these are shown for

ENCODING PICTURES AND WORDS each trial, yielding eight unique trial types (e.g., hippo-mouse and mouse-hippo are two different trial types). The eight trial types map into four relevant values or solutions (e.g., in a ++ problem, hippo-mouse and mouse-hippo will both yield the solution ANIMAL), which in turn map into two dimensions (e.g., the relevant values ANIMAL and VEHICLE belong to the category dimension). The relevant dimension is the dimension which contains the relevant value for a particular trial type within a particular inference condition. In Experiment 1, we replicated Van Rijn's (1973) study with semantic stimuli presented as either pictures or words. We included all four inference conditions, ++, -1—, —+, and , to insure that the change from geometric designs to semantic stimuli did not affect the order of problem difficulty. However, we were mainly interested in the ++ and conditions, in which the likely processing strategies are clearer. We compared reaction times for both pictures and words across inference conditions in an effort to show that stimulus identification, as well as the abstraction of information from the stimuli, does not interact with subsequent processing stages. Experiment 1 Method Subjects and design. The 48 subjects who participated in the experiment each received $1.50. They were randomly assigned in equal numbers to one of eight conditions generated by the factorial combination of problem (++, , —h, and H—) and stimulus type (pictures or words). Stimuli. A 2 X 2 matrix was constructed from attribute (LARGE or SMALL) and category (ANIMAL or VEHICLE) dimensions, and instances were found which exemplified the combination of relevant values for a given cell. Hippo was used for LARGE ANIMAL, mouse for SMALL ANIMAL, bus

for LARGE VEHICLE, and car for SMALL VEHICLE. Thus, there were four trial types in which a category was common to both instances (e.g., hippomouse, mouse-hippo, bus-car, and car-bus), and four in which an attribute was common to both instances (e.g., hippo-bus, bus-hippo, mouse-car, and car-mouse). None of the solutions were physically present in the instances. For example, there was nothing in the configuration of either the word or the picture (an outline drawing) hippo which demonstrated that a hippo is a LARGE ANIMAL. This is a fact that the subject must

175

know or determine from the configuration presented in order to perform the task. Procedure. Van Rijn's (1973) procedure was followed as closely as possible. The subjects were seated in front of a rear projection screen. Under the screen were two lights which indicated the assignment of the stimulus (i.e., whether or not it contained the relevant value). After the nature of the problems and solutions were described to the subject, he had 8 practice problems, followed by 72 working problems. The first block of 8 working problems was not used in the data analysis, so each subject yielded a total of 64 data points. The eight trial types were randomized for each block of 8 trials. A trial consisted of the following sequence of events: (a) The subject pushed a start button; (b) after a delay of about .5 sec, the first slide came on, causing a photocell to activate a millisecond timer, (c) When the subject pressed the button again, the timer was stopped, the second slide (from a different projector) replaced the first slide instantaneously, and a second timer was started, (d) The subject spoke his response into a Grason-Stadler voice key, causing the second timer to stop, (e) The experimenter recorded the two times and reset the apparatus for the next trial.

Results and Discussion The rejection region for all of the following analyses is p < .05, and all MSes are reported in milliseconds. The mean of both the first and second reaction times for each block of 8 trials was calculated for every subject, with error data omitted. Table 1 shows the mean reaction times and error rates for the four inference conditions; each mean is based on 192 scores (2 reaction times per block X 8 blocks per subject X 12 subjects per condition). The data from Van Rijn's study, using geometric stimuli, are included for comparison. Planned orthogonal contrasts showed that we replicated Van Rijn's essential results with semantic stimuli; the ++ inferences took less time than the inferences, F(l, 44) = 15.56; the + — and — + inferences were not reliably different from each other, F(l, 44) < 1; and the average reaction times to solve -f + and inferences were reliably faster than those to solve +— and — + inferences, F(l, 44) = 17.96, Af.?, = 359,281.53. It is likely that the subjects in the latter two conditions used a much more heterogeneous set of strategies than did the subjects in the ++ and conditions; in any case, for present purposes,

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

176

TABLE 1 MEAN REACTION TIMES AND PERCENT ERROR IN THE FOUR INFERENCE CONDITIONS WITH EITHER SEMANTIC (EXPERIMENT 1) OR GEOMETRIC (VAN RIJN, 1973) SOLUTIONS Stimulus type Semantic

M

Geometric

M

% error % error

606.95 .78 593.73 1,4

848.27 2.6 731.44 2.5

897.12 2.9 808.84 3.1

924.75 2.5 780.04 7.0

we intend to restrict the discussion to and conditions only. The + + and data were analyzed in a mixed design, with problem and stimulus type (pictures versus words) as the 'betweensubject s variables and reaction time (first or second) and blocks (1 to 8) as the withinsubjects variables. The main effect of problem, P(\, 20) =23.83, MSe = 234,605.49, was expected on the basis of planned comparison. The main effect of reaction time, F(l, 20) = 191.12, MSe = 156,130.94, indicates that first reaction times were always faster than second reaction times (449 versus 1006 msec, respectively). This is to be expected because in theory, the first reaction time reflects processing of the first stimulus, while the second reaction time covers both encoding of the second stimulus and any subsequent problem solving activities. As required by a literal levels-of-processing interpretation, the effect of stimulus type was not reliable, F(l, 20) = 3.59, .05 < p < .10. The difference is large enough to give us some concern, however (the average second reaction times for pictures and words were 950 and 1062 msec, respectively). Importantly, there were no reliable interactions with stimulus type. Both the main effect of blocks, F~(7,140) = 7.95, MSe = 7,436.15, and the Blocks X Reaction Time interaction, P(7, 140) = 5.54, MS. = 3,261.65, suggested that practice might have either enhanced or diminished the picture-word difference, so we repeated the foregoing analysis on data from the last block of trials only. There were again significant effects of problem, P(l, 20) =22.88, MSe - 28,041.91, and reaction time, F(l, 20) = 161.21, MSe = 27,378.13. In addition, a reliable main

effect of stimulus type, -F(l, 20) = 4.60, MSe = 28,041.91, materialized, although once again stimulus type did not interact with any other variable. Table 2 shows the mean second reaction times for all stimulus type and problem conditions from both analyses. It can be seen that by the last block of trials, the relative difference between + + and —>• is virtually identical for pictures and words (280 versus 271 msec) and pictures are on the average 136 msec faster than words. Since the ++ and difference is constant across;materials, we can unambiguously assign the picture-word difference to the stimulus encoding stage. Regardless of the type of inference, words require 136 msec more than pictures before their relevant semantic characteristics are available to be used in an inference. The data are thus in accord with a levels-of-processing analysis which assumes that pictures are identified faster than words on the basis of physical discriminability, but once identified, extracting semantic information requires the same amount of time for both stimulus types. To get additional evidence for this argument, the data from the last block of trials were partitioned into trials which had either category (e.g., ANIMAL or VEHICLE) or attribute (e.g., LARGE or SMALL) relevant values. Recent dual-coding approaches (e.g., Paivio, Note 1) assume that superordinate categories may be represented via prototypical exemplars; on the other hand, information about attributes of specific instances, especially visual attributes such as size or color, may easily be stored within the imaginal system (Moyer, 1973; Paivio, TABLE 2 MEAN REACTION TIMES FOR PICTURE AND WORD STIMULI AS A FUNCTION OF THE TYPE OF INFERENCE Experiment 1 (Second reaction time) Stimulus type

All blocks ++

--

Pictures Words

797 925

1,104 1,200

Words Pictures:

128

96

Experiment 2

Block 8

Block 8

++ --

++

787 927

1,067 1,198

819 1,077 967 1,206

140

131

148

--

129

ENCODING PICTURES AND WORDS 1975). Since car was the only one of the four instances which might be considered prototypical of its category, a dual-coding approach should predict that for picture stimuli, trials with the size dimension relevant should be responded to faster than trials with the category dimension relevant, while for word stimuli, there should be no difference between the dimensions. Accordingly, the partitioned data were analyzed in a mixed design with problem and stimulus type the between-subjects variables and reaction time and relevant dimension the within-subjects variables. The main effect of relevant dimension, F(\, 20) = 7.12, and the Relevant Dimension X Problem interaction, F(l, 20) = 8.36, MSe - 13,334.93, were the only new reliable effects. Table 3 shows the second reaction times for the interaction. The dualcoding predictions were not supported; there is virtually no difference between category and attribute trials in the ++ condition for either pictures or words. In the condi-, tion, however, the category dimension took longer than the attribute dimension, and the difference was similar for both pictures (148 msec) and words (195 msec). Since the category-attribute difference was roughly the same across stimulus materials, but interacted with problems, the locus of the effect seems likely to be in the problem solving stage. One possible explanation derives from the fact that the attribute values we used were linguistic opposites, while the category values were not; this would facilitate the proposed complement-taking operation for subjects but would not be expected to affect the ++ subjects. The results of Experiment 1 can be summarized as follows. First, since the relative reaction time difference between ++ and was constant across stimulus materials, with solution characteristics held constant and explicitly known, we can be sure that stimulus encoding does not interact with problem solving operations and that the picture—word difference in speeded inference resides in the stimulus encoding stage. Second, the relevant dimension interacted with inference type independent of stimulus modality, and we can therefore tentatively assign the category-attribute difference to

177

TABLE 3 MEAN REACTION TIMES FOR CATEGORY AND ATTRIBUTE RELEVANT DIMENSIONS AS A FUNCTION OF STIMULUS TYPE AND INFERENCE FOR EXPERIMENT 1 (BLOCK 8 DATA) Relevant dimension

Category Attribute Category — Attribute:

Pictures

Words

777 800

1,140 992

917 937

1,303 1,108

-23

148

-20

195

the problem solving stage. Both of these results lend support to a levels-of-processing interpretation of stimulus encoding, in which encoding involves both tacit identification of the stimulus and abstraction of characteristics which will be relevant to the task at hand. Despite the fact that the present task explicitly required abstracting the same semantic characteristics from both pictures and words, there was a small but significant difference in the encoding time in favor of pictures. This suggests that there were differences in discriminability among the pictures which allowed them to be identified from among one another sooner than the words could be. Experiment 2 The purpose of Experiment 2 was threefold : First, we wanted to replicate the picture-word effect of Experiment 1. Second, the inferences we made about the data from Experiment 1 were based on the assumption that both stimulus encoding and problem solving operations were measured by the second reaction time (time from the onset of the second stimulus). To obtain a cleaner and more rigorous measure of these effects, we held the duration of the first stimulus constant in Experiment 2 and measured reaction time from the onset of the second stimulus. Third, we have argued that the picture-word effects in a speeded-inference task reside in the stimulus encoding stage. But since all subjects responded verbally, subjects who received word stimuli might have experienced a form of response interference, which would confound this inter-

178

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

pretation. We tested this by replicating the + + and conditions of Experiment 1, having half the subjects respond verbally and half respond by pressing buttons. The picture-word difference should vanish or reverse itself in the button group if response interference was responsible for it. Method Subjects and design. The subjects were 40 undergraduates, who participated for course credit; they were randomly assigned to one of eight groups formed by the factorial combination of problem (H-+ or ), stimulus type (pictures or words), and response mode (voice or buttons). Procedure. There were two changes in the procedure from Experiment 1. First, the "stimulus information" lights were eliminated because in Experiment 1 they were ignored by the subjects after the first few trials. Second, the interval from the onset of the first stimulus to the onset of the second stimulus was fixed at 750 msec for all subjects. Reaction time was measured from the onset of the second stimulus until the subject's response. The instructions were changed to accommodate these differences. For the button group, four telegraph keys were arranged in a semicircle, approximately 2 in. from the start button and equidistant from each other. Each key was labeled with one of the four relevant values. The subjects in this group werfe instructed to keep their index finger on the start button until they were ready to respond. In all other respects, the procedures of Experiment 1 and 2 were identical, as were the stimulus materials.

Results and Discussion The overall error rates were 2.8%, 3.1%, 2.0%, and 3.8% for pictures ( + + and ) and words ( + + and ), respectively. For comparability to Experiment 1, the mean score for the last block of 8 trials was calculated for every subject, with error data omitted. These data were analyzed in a between-subjects design with stimulus type, problem, and response mode as the factors. The error term for all of the effects was MSK = 28,478.99. The main effect of stimulus type was reliable, F(l, 32) = 6.76, showing that pictures were again faster than words (948 versus 1,087 msec). This is a difference of 139 msec, which is almost identical to the difference (136 msec) found for the second reaction time data in the last block of 8 trials in Experiment 1. The main effect of problem showed that the ++ infer-

ences were reliably faster than the inferences, F(l, 32) = 21.67, and the main effect of response mode showed that it was easier to respond vocally (951 msec) than by pushing buttons (1,084 msec), F(\, 32) = 6.19. None of the interactions approached significance. Table 2 shows the Stimulus Type X Problem data for the last block of trials from the present experiment, where it can be compared with the data from Experiment 1. It can be seen that in the present experiment, the difference between inference conditions is again comparable across stimulus materials (258 versus 239 msec for pictures and words, respectively). Furthermore, the fact that the reaction times are of the same order of magnitude lends support to our hypothesis about the second reaction times of Experiment 1. Lastly, since the picture-word difference held up across response mode and was in fact even somewhat larger in the button group than in the voice group >(163 versus 114 msec, respectively), we can rule out response interference as the cause of this difference. Experiment 3 When the data from our first experiment were partitioned into trials which required category responses (e.g., hippo-mouse: ANIMAL) and trials which required attribute responses (e.g., hippo-bus: LARGE), there was no difference in reaction times for the + + problems, but the category trials were reliably slower than the attribute trials on problems. Since this effect was of the same magnitude for both pictures and words, it was tentatively assigned to the problem solving stage of the task. We proposed that finding the complement of the common value was facilitated on inferences that required attribute responses because the attribute values we used (LARGE-SMALL) were linguistic opposites, while the category names (ANIMAL-VEHICLE) were not. Experiment 3 was designed to test this hypothesis. We constructed four conditions. Each involved two different stimulus matrices in which the relevant values of one dimension were two categories and the relevant values of the other dimension were two attributes. In the Both condition, the values of both

ENCODING PICTURES AND WORDS

the category and attribute dimensions were linguistic opposites; in the Attribute-opposite condition, only the relevant values of the attribute dimension were linguistic opposites. Similarly, in the Category-opposite condition, only the category dimension had linguistically opposite relevant values, and finally, neither dimension had linguistically opposite relevant values in the None condition. According to our reasoning about the properties of the task, it will be inappropriate to make comparisons across conditions, since the different stimulus matrices involved may require different amounts of analysis to completely encode their particular instances. However, we may compare the time required to find the complement of the common value for the category and attribute dimensions within each of the four conditions, to show whether linguistic "oppositeness" was the factor underlying the original category-attribute difference. More specifically, we have hypothesized that the "timeto-complement" the common value is the only operation which distinguishes from + + inferences; this time may be estimated for each dimension within a condition by subtracting the mean ++ from the mean reaction times for that dimension. If our hypothesis is correct, the time-to-complement category and attribute values should be equal in the Both and None conditions, since linguistic oppositeness is either present or absent from both dimensions. On the other hand, in the Attribute-opposite condition, the time-to-complement attributes should be shorter than the time-to-complement categories, and in the Category-opposite condition, the reverse should be true. This is the same as saying that for these conditions, a property of a particular relevant dimension interacts with inference type during the inference or problem solving process and not during encoding. Method Subjects and design. The subjects were 48 undergraduates, who participated in the experiment for course credit. The first 36 subjects were randomly assigned to one of the conditions formed by the factorial combination of problem (++ and ) and the number and type of dimensions whose values were linguistic opposites (Both, None,

179

Category opposite, and Attribute opposite). Our initially constructed Category-opposite stimulus matrices were extremely difficult, and because of the time required to get new stimulus materials, all 12 subjects in the Category-opposite condition were tested during the final week of the semester. For these reasons, 12 additional subjects were run during the following semester as a replication of the Category-opposite condition. There were two different stimulus matrices for each condition; half the subjects received Matrix A first and half received Matrix B first. Instances were found for the cells of each matrix, again yielding eight trial types; four in which the category of the instances was the same (e,g., hippo-mouse, mouse-hippo: ANIMAL) and four in which the attribute was the same (e.g., hippo-bus, bus-hippo: LARGE). There were eight trials in a block, one of each trial type in a random order. All of the stimuli were presented as words. The stimulus materials are shown in Table 4. Procedure. The subjects received one block of practice and then four more blocks of trials on each stimulus matrix in their condition. Otherwise, the procedure was identical to the voice group of Experiment 2.

Results and Discussion The error rates were 2.1%, 3.4%, 2.0%, and 4.3% for Both, None, Attribute-opposite, and Category-opposite conditions, respectively. Within each condition, a mean reaction time was obtained for + + and inferences in which either an attribute or a category was the relevant value, and the difference between these means (i.e., the time-to-complement that dimension) was obtained. These data are shown in Table 5. The difference between the time-to-complement attributes versus categories was tested against the standard error of the difference between both inference types within a condition, combined across relevant dimensions. As predicted, the attribute dimension was "easier" than the category dimension only in the Attribute-opposite condition, £(22) = 2.58, and the category dimension was easier than the attribute dimension in the Category-opposite condition, £(22) = 4.00, and in the replication of the Category-opposite condition, £(22) = 4.82. There was no difference in complementtaking times between the dimensions in either the Both or None conditions. In general, these results confirm the hypothesis that subjects making inferences locate the common value and find its com-

180

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR. TABLE 4 STIMULUS MATERIALS FOR EXPERIMENT 3 Condition

Matrix A SOLID

WHITE

Both

BLACK

salt coal

Matrix B LIQUID

milk coffee

MALE YOUNG OLD

boy father ANIMAL

None

RED YELLOW

apple banana

robin canary

SMALL FAST

CLOTHING

Attribute opposite

Category opposite

LONG SHORT

BLACK COLD

hour minute

gown skirt

SOLID

LIQUID

coal ice

plement and that if a dimension has relevant values that are linguistic opposites, this process is facilitated. General Discussion Taken together, the results of the first two experiments rule out a strict levels-ofprocessing interpretation, which would claim that pictures should have been equal to words simply on the grounds that the same depth of information was required from both. They also seem inconsistent with a dual-coding hypothesis, however, since words, which should have more direct access to semantic or generic information, were slower than pictures. They are consonant with an interpretation which says that the same semantic information should in fact be equally derivable from both pictures and words, but that pictures may contain physical information which makes them more discriminably different from one another (more readily identifiable) than are words. From this viewpoint, the picture-word difference in speeded inference is an encoding phenomenon which results from differences in identifiability. In other words, within a levels-of-processing framework, one may expect reaction time

coffee beer

LARGE SMALL

girl mother BOAT

mouse cheetah

canoe cruiser

ANIMAL

VEHICLE

hippo mouse FRUIT

YELLOW SMALL

FEMALE

banana grape

bus car VEGETABLE

squash pea

differences in favor of pictures when performance requires discriminating a stimulus from among a set of other possibilities before proceeding with the task. An example of this is the increase in reaction time obtained when the probe stimulus in a memory search task is visually degraded (Sternberg, 1969), since the only encoding operation necessary for successful performance is the identification (or perhaps recognition) of the stimulus. The speeded-inference task requires further stimulus analysis in order to proceed to subsequent rionencoding operations. It remains to be seen whether we can use this task to get reaction time differences in the abstraction of different depths of information within pictures and words. The general prediction is that with discriminability held constant, semantic or conceptual solution-relevant information should require longer encoding times than should physical information, regardless of stimulus modality. LEVELS OF PROCESSING WITHIN PICTURES AND WORDS Experiment 4 In the present study, we attempted to demonstrate an artificial picture-word dif-

ENCODING PICTURES AND WORDS

ference, using only picture stimuli, by varying whether the relevant values were physically present in the pictures. In other words, we tried to manipulate the depth of processing which would be required to abstract a relevant characteristic from pictorial stimuli for a speeded inference. Characteristics physically present in the stimulus should become available to the subject sooner than characteristics which are not physically present, but once available, should not influence subsequent processing stages. This results in the prediction that with the same picture stimuli, inferences with "perceptual" solutions will be faster than inferences with "conceptual" solutions, and the facilitation will be confined to the stimulus encoding stage'of the task. Method Subjects, The subjects were 40 undergraduates, who participated for course credit, and were divided equally among the four experimental groups. Materials and design. In an attempt to avoid confounding frequency or familiarity with depth of processing, we used the nonsense creatures pictured in Figure 2 as stimulus instances. Four of the eight possible trial types which can be created from these instances have either HORNS or TAILS as their solution. Since HORNS and TAILS are pictured in the instances, this dimension was a perceptual dimension for all subjects. The only difference between the perceptual and conceptual conditions was in what the subjects were to call the relevant values of the four remaining trial

181

types. In the perceptual condition, the subjects were told that ROUND and SQUARE were the other relevant values; again, both of these values are physical properties of the instances. In the conceptual condition, the subjects were told that ALIVE and EXTINCT were the other relevant values; neither of these are perceptually inherent in the instances, and this is therefore a conceptual dimension. In other words, given a ROUND creature with HORNS followed by a ROUND creature with a TAIL, a perceptual subject making a ++ inference would respond round, while a conceptual subject would respond extinct. Given a ROUND creature with HORNS followed by a SQUARE creature with HORNS, both perceptual and conceptual subjects would respond horns. We are particularly interested in comparing ALIVE-EXTINCT trials to HORNS-TAILS trials within the conceptual condition, since this would enable us to see a depth effect within individual subjects. However, we predicted that all of the trials with perceptual relevant values (i.e., HORNS, TAILS, ROUND, and SQUARE) would yield relatively fast reaction times, while the ALIVE-EXTINCT trials would be much slower. Procedure. Each subject was instructed with a picture similar to Figure 2, with the labels on the picture appropriate for his condition. The subjects were told about four nonsense creatures that had either HORNS or TAILS and were of one of two types (ROUND-SQUARE or ALIVE-EXTINCT, depending on the group). The remainder of the procedure was identical to the voice condition of Experiment 2.

Results and Discussion The error rates were .80%, 1.4%, 2.3%, and 3.8% for the perceptual ( + + and ) and conceptual ( + + and ) conditions, respectively. The data from the last block

TABLE S MEAN REACTION TIMES FOR + + AND INFERENCES AND THE TIME-TO-COMPLEMENT THE RELEVANT VALUE AS A FUNCTION OF THE TYPE OF RELEVANT DIMENSION AND WHETHER ITS VALUES WERE LINGUISTIC OPPOSITES Relevant dimension

Inference

Attribute __ I. Time-to-complement common attribute: Category __ II. Time-to-complement common category: Difference between complementing times [(I) - (II)]:

Attribute opposite

Category opposite

Category-opposite replication

Both

None

881 1,236

1,151 2,424

1,088 1,862

919 1,298

933 1,378

35S

1,273

774

379

445

853 1,364

1,318 1,942

1,225 1,588

1,012 1,421

938 1,315

511

624

363

409

377

-156" (60.45)

649" (162.33)

411" (85.39)

Note. Numbers in parentheses are standard errors of the difference between ++ and • p < .05.

-30 (69.31)

68 (58.21)

across both relevant dimensions.

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

182

TABLE 6 MEAN REACTION TIMES TO MAKE ++ AND INFERENCES AS A FUNCTION OF THE DEPTH OF PROCESSING REQUIRED BY THE RELEVANT DIMENSION FOR EXPERIMENT 4 (BLOCK 8 DATA) Perceptual condition Inference

Conceptual condition

RoundSquare

HornsTails

Horns- AliveTails Extinct

732 1,017

778 1,041

679 955

1,070 1,305

285

263

276

235

of problems for each subject were partitioned into the two main dimension types for his condition (i.e., HORNS-TAILS and either ROUND-SQUARE or ALIVE-EXTINCT), and an overall mean was calculated for each dimension, with error data omitted. These data were analyzed in a mixed design with condition (perceptual or conceptual) and problem ( + + or ) as the between-subjects variables and dimension as the within-subjects variable. Dimension 1 was HORNSTAILS and Dimension 2 was either ROUNDSQUARE or ALIVE-EXTINCT. Note that by combining the ROUND-SQUARE and ALIVE-EXTINCT times, we were decreasing the difference between perceptual and conceptual conditions. In spite of this, the main effects of condition, F(l, 36) = 6.68, problem, F(l, 36) = 38.51, MSe = 36,416.01, and dimension, F(l, 36) = 24.42, as well as the interaction of condition and dimension, F(l, 36) = 35.33, MSe = 23,169.46, were all statistically reliable. All of the effects except problem reflected the longer reaction times on trials when either ALIVE or EXTINCT was the relevant value in the conceptual condition (see Table 6). Moreover, within this condition, the reaction times on trials with HORNS or TAILS relevant were comparable to those of the perceptual condition as a whole. As further testimony to the robustness of this effect, by the last block of trials, 18 out of 20 subjects in the conceptual condition (both ++ and ) had longer reaction times on their ALIVE-EXTINCT trials than on their HORNS-TAILS trials, with a range from 39-822 msec and a mean difference of 424 msec. As a contrast, in the per-

ceptual condition, 13 subjects had longer reaction times for HORNS-TAILS trials and 7 had longer reaction times for ROUND-SQUARE trials; for these 20 subjects combined, the range of the difference scores was from 5287 msec, with a mean difference of 103 msec. Since the reaction time difference between + + and inferences is again roughly similar for both dimensions within and across conditions (and comparable to the difference obtained in the first two experiments), we know that the dimension effects are an encoding phenomenon. Furthermore, since all subjects saw exactly the same stimuli, the reaction time differences among dimensions cannot be attributed to discriminability, but must be taken as a genuine levels-ofprocessing effect. Experiment 5 In the previous experiment, we obtained depth effects with pictures by having some of the relevant values be redundant with physical features of the stimuli. If we could show an analogous effect with word stimuli, it would provide converging evidence for the notion that at least some picture-word differences can be attributed to the subjects' ability to use discriminating physical features in pictures which are not usually availExtinct Round

— Alive or — Square

Horns

Tails

FIGURE 2. Nonsense creatures used as stimuli in Experiment 4. (The subjects- were instructed with the use of a photograph similar to this figure except that only the values appropriate to their condition were displayed.)

ENCODING PICTURES AND WORDS able in words. The more that task-relevant processes can take advantage of physical attributes (e.g., in a visual search task [Paivio & Begg, 1974]), the more we should expect a picture-word difference. In the speeded-inference task, we can construct word matrices in which some of the relevant values are redundant with physical features of the referents of the words. For example, bananas and canaries, as objects, both share the property yellow, while apples and robins share the property red. Color is thus perceptually inherent in the referents and conceptually inherent in the names of the referents. However, if we print the stimulus words in yellow or red ink, then color becomes a physical feature of the words as well. "Normal" stimuli would presumably have to undergo more analysis than colored stimuli in order to derive the colors, in view of the fact that physical information should be available sooner than semantic information. Note that the subjects who receive colored stimuli cannot ignore the meaning of the words because a random half of their trials require category answers (i.e., FRUIT or BIRD) that the colors cannot cue. Thus, the important prediction from a levels-ofprocessing point of view is that RED-YELLOW trials will yield faster inferences when the stimulus instances are printed in color than when they are printed in black ink, and FRUIT-BIRD trials should not differ as a function of stimulus materials. The only real difference between Experiment 4 and the present experiment is in the stimulus materials ; the logic is the same, insofar as trials with physical relevant values should yield faster inferences than trials with conceptual relevant values. Method Subjects. The subjects were 36 undergraduates, who participated for course credit, and were randomly assigned to one of four experimental groups obtained by combining problem (++ or ) and group (C-N or N-C). Materials and design. The stimulus matrix used for the present experiment was the same as Matrix A for the None condition in Experiment 3 (see Table 4). Half the subjects saw their instances printed in the appropriate color (either RED or YELLOW) and the other half saw them in black and white.

183

The subjects were instructed with index cards that showed the stimulus matrix appropriate for their group. All subjects received two practice and eight experimental blocks of trials with their original stimulus type (color or normal) ; a block consisted of one of each of the eight trial types in a random order (four with RED or YELLOW relevant and four with FRUIT or BIRD relevant). After this, they were transferred to the other stimulus type and run for four additional blocks, so that in a crude sense, each subject served as his own control. Group C-N (color-normal) received colored stimuli first and were transferred to normal stimuli, and Group N-C (normal-color) received normal and then colored stimuli. The eighth block of original trials was identical to the fourth block of transfer trials. The remainder of the procedure was the same as in Experiment 4.

Results and Discussion The error rates were 1.7%, 2.7%, 1.1%, and 2.9% for Group C-N ( + + and ) and Group N-C ( + + and ), respectively. The data from the last block of both original and transfer trials were partitioned into trials on which either FRUIT-BIRD or RED-YELLOW values were relevant, and a mean reaction time was calculated for each dimension for every subject, excluding errors. These were analyzed in a mixed design, with relevant dimension and stage (original or transfer trials) as within-subjects variables and group (C-N or N-C) and problem ( + + or ) as between-subjects variables. The main effect of problem was reliable, as expected, F(l, 32) = 151.15, MSe = 42,167.58. The main effect of stage F(l, 32) =4.17, MSe = 18,572.95, showed that the transfer trials were slower in general than the original trials (1,031 versus 985 msec), probably because the subjects had 10 blocks (including practice) with their original stimulus type and only 4 transfer blocks. The main effect of relevant dimension, F(l, 32) = 25.88, MS* - 21,493.06, showed that the RED-YELLOW dimension was 124 msec faster than the FRUIT-BIRD dimension. Since the reaction time difference between + + and inferences was roughly the same for both groups (414 versus 428 msec for C-N and N-C, respectively) and close to the difference obtained for this problem in Experiment 3 (411 msec), we can again assign differences across groups to encoding operations alone.

184

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

TABLE 7 MEAN REACTION TIMES FOR TRIALS WITH COLORED AND NORMAL STIMULI AS A FUNCTION OF THE RELEVANT DIMENSION AND STAGE OF THE EXPERIMENT FOR EXPERIMENT S (BLOCK 8 ORIGINAL TRIALS AND BLOCK 4 TRANSFER TRIALS) KRUIT or BIRD relevant Group

Color-normal Normal-color Normal — Color:

RED or YELLOW relevant

Original words

Transferred to

Original words

Transferred to

color : 1,026 normal : 1,081

normal : 1,127 color : 1,045

color : 788 normal : 1,045

normal : 1,093 color: 857

55

82

257

236

The most important results for the present hypothesis are the interactions that would be expected from fast reaction times on trials when either RED or YELLOW is the relevant value and the stimulus instances are printed in color, regardless of whether those trials were in the original or transfer stage. The Group X Stage effect, F(l, 32) = 48.29, MSe = 18,572.95, showed that the colored stimuli were faster than the normal stimuli on the original trials (907 versus 1,063 msec) and on the transfer trials (951 versus 1,111 msec). The Group X Stage X Problem interaction, F(\, 32) = 8.01, MS* 18,572.95, showed that this relationship held for both ++ and inferences; and most importantly, the Group X Stage X Relevant Dimension interaction, F(l, 32) = 16.92, MSf. = 16,672.92, showed that the advantage for the colored stimuli was limited to trials when either RED or YELLOW was relevant, as predicted. Thus, the subjects who originally saw colored words (i.e., the Group C-N) made faster inferences on RED-YELLOW trials than on FRUIT-BIRD trials; when transferred to words in black ink, reaction time did not differ as a function of dimension. The reverse pattern was obtained for subjects in the Group N-C. Table 7 shows the data from this interaction. We repeated the foregoing analysis on the RED-YELLOW trials alone in order to make sure that the predicted interaction (i.e., Group X Stage) is reliable for these problems. The results were essentially the same; the main effect of problem, F(l, 32) = 81.56, MSe = 36,344.27, the Group X Stage interaction, F(l, 32) = 45.24, and the triple interaction, F(l, 32) = 7.84, MSe = 24,148.50, were all reliable. Table 8 shows the triple interaction.

This experiment demonstrates, once again, that physical information is generally available sooner than conceptual or semantic information, and we should expect faster reaction times when the relevant dimension is physically present in the instances (as in Experiments 4 and 5) or when the physical characteristics of the instances allow them to have a "head start" in the encoding process (as in Experiments 1 and 2). While words are visual stimuli with perceptible physical features, the physical properties of words are not usable under most circumstances. Experiment 6 The results so far suggest that depth effects can be obtained within picture and word stimuli, but that when pictures and words are both to be analyzed to the same depth (e.g., Experiments 1 and 2), pictures generally have the advantage of better discriminability, giving them a head start. It is reasonable, at this point, to wonder whether- there can ever be a situation in which no picture-word difference occurs, as was originally expected. The relationship between the name of an object and a picture which represents that object is usually one-to-many; that is, a single word is usually applicable to an almost infinite set of pictures which are referents of that word. Shapes of the United States (or countries and perhaps other stimuli) constitute a set of pictures which have the unique property of a one-to-one relationship between shape and name. Shapes of states may thus function at the same level of abstraction (i.e., the generic) as do names. If the shape of a state can serve as a symbol, it may be as identifiable as its name, leading

ENCODING PICTURES AND WORDS to an expectation of no reaction time difference between shape and name. Alternatively, shapes of states may be experientially "rare" enough that their advantage in discriminability will be offset by the speed with which subjects can read their names. In the following experiment, we used names or shapes of states as stimuli and held responses constant at a relatively deep level in an attempt to demonstrate equal accessibility of conceptual information across these materials. Method Subjects and design. The subjects were forty undergraduates, who participated for course credit, and were assigned to one of the four groups made by the factorial combination of problem (+4- or ) and stimulus type (names or shapes). Except for changing the instruction card, the procedure was identical to the last two experiments. Materials. Four of the United States—New York, Florida, Alaska, and Arizona—which can be uniquely characterized as being north or south and east or west were chosen. Two of the states were presented on each trial. Half the subjects in each inference condition saw the names of the states as stimuli, and half saw the outlines of their shapes. Thus, for example, a subject in the ++ condition who saw either the names or the shapes of New York and Alaska would respond with the location north. We felt that the geographical location of a state should be no more perceptually a part of its shape than of its name. That is, the location should be an attribute which is equally derivable from both name and shape, since it is inherent in the meaning or concept of the state. There are again eight unique trial types, two for each geographical location (e.g., New YorkAlaska, Alaska-New York: NORTH). A block of trials consisted of all eight types in a random order. On any given trial, a subject was shown two states and had to respond verbally and as rapidly as possible with the location they had in common (if he was in the +4- condition), or its opposite (if he was in the condition).

185

Results and Discussion The error rates were 2.5%, 3.0%, 2.7%, and 3.2% for shapes ( + + and ) and names ( + + and ), respectively. A mean reaction time was calculated for the last block of problems for each subject, with error data omitted. The data were then analyzed in a 2 X 2 design, with problem and stimulus type the between-subjects variables. Only the main effect of problem, F(l, 36) = 9.34, MSe = 53,721.96, was reliable. The mean reaction times for the ++ and problems were 973 and 1197 msec, respectively. While the absence of a main effect of stimulus type (the mean reaction times to shape and name stimuli were 1027 and 1144 msec) is exactly what we predicted if the geographical location of a state was indeed the sort of information that required a deep level of processing to obtain and if shapes and names were functioning at the same level of abstraction, the within-subject variability was large enough to cause concern about these conclusions. When we calculated the mean score for each subject, for example, the difference in reaction time between the fastest and slowest subjects in each group was 634, 415, 926, and 852 msec for shapes ++, names ++, shapes , and names , respectively. Each subject seemed to reach a level of proficiency that might have been determined by some combination of his inference condition and stimulus type, but given that these were betweensubjects variables, there was no way to be certain. We felt that a within-subject comparison of performance to name and shape stimuli would provide a stronger demon-

TABLE 8 MEAN REACTION TIMES FOR TRIALS WITH "Run" OR "YELLOW" RELEVANT AS A FUNCTION OF COLORED OR NORMAL STIMULI, TYPE OF INFERENCE, AND THE STAGE OF THE EXPERIMENT FOR EXPERIMENT 5 (BLOCK 8 ORIGINAL TRIALS AND BLOCK 4 TRANSFER TRIALS) + + Inferences Group

Color-normal Normal-color Normal — Color:

Inferences

Original words

Transferred to

Original words

Transferred to

color: 638 normal: 798

normal: 831 color: 704

color: 938 normal: 1,292

normal: 1,355 color: 1,011

160

127

354

344

186

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

Results and Discussion

subjects variable and stimulus type, days, and blocks the within-subjects variables. Only the main effect of days was significant, F(2, 4) = 25.56, MS. = 19,705.53. However, since there were only two subjects per inference condition, both main effects of problem and stimulus type were tested against only two degrees of freedom for error. We therefore decided to test both these effects nonparametrically via the MannWhitney [/-test. A mean reaction time was calculated for each subject for each day, yielding 6 scores per subject (3 for names and 3 for shapes) and a total of 12 scores per inference condition. When the reaction times were rank ordered according to inference condition, then U(12, 12) = 136, which was statistically reliable. When they were rank ordered according to stimulus type, then U(12, 12) =80, which was not significant. The mean reaction times to name and shape stimuli were nearly identical; in the ++ condition, the average reaction time for name and shape problems, respectively, was 748 and 750 msec; in the condition, the average reaction time was 1,069 and 1,067 msec. Several conclusions about the results seem to be warranted by an inspection of the left side of Figure 3, which shows the relationship between stimulus type and problem over days. The reaction time performance of the subjects improved across days and was not a function of stimulus type. Moreover, the results support a levels-of-processing interpretation to the extent that the same depth of conceptual information was equally derivable from both pictures and words. They also appear to support the hypothesis that shapes of states function as symbols by virtue of their one-to-one relationship to a name.

The error rates, averaged across subjects, were 1.0%, 3.9%, 1.6%, and 4A% for shape ( + + and ) and name ( + + and ) conditions, respectively. A mean score was calculated for each of the last eight blocks of each day. Thus, each subject yielded 48 data points; 24 for name stimuli and 24 for shape stimuli. The data were analyzed in a mixed design, with problem the between-

Experiment 8 If the shapes of states really do function as symbols, then it should be possible to perform a manipulation analogous to what we did in Experiment 5, when we obtained faster performance with word stimuli whose built-in perceptual features were redundant with the responses. We should be able to add features to the shapes of states which

stration that these two stimulus types provide equally effective access to solution-relevant information.

Experiment 7 As in Experiment 6, the purpose of Experiment 7 was to show that comparable performance levels obtain with names and shapes when the information required by the task resides in the meaning rather than in the physical attributes of those stimuli. We gave four subjects extended practice on speeded inference with the same materials used in Experiment 6, with all the trials of a given stimulus type being blocked, so that if it was possible to develop a rapid, perceptual-processing strategy with the shape stimuli, the subjects could do so. We predicted a steady decrease in reaction time across days, regardless of stimulus type, until asymptotic performance was reached. Stimulus type should not be a significant factor in determining a subject's final level of proficiency. Method Subjects and design. The task and materials were identical to those used in Experiment 6. Three volunteers and the first author had 10 blocks of trials with eight trials per block for 6 days. There were two subjects (one male and one female) in each inference condition. The stimulus type (names or shapes) was blocked by days such that during the first 3 days, one subject in each inference condition had name stimuli, while the other subject had shape stimuli; during the next 3 days, each subject was switched to the other stimulus type, after being appropriately instructed. In all other respects, the procedure was identical to that in Experiment 6.

ENCODING PICTURES AND WORDS

1100

187



500 DAY1

DAY 2 EXPERIMENT 7

DAY 3

DAY 1

DAY 2 EXPERIMENTS

DAY 3

FIGURE 3. The Stimulus Type X Problem data for Experiments 7 and 8 as a function of days.

would allow them to be identified or discriminated from each other sooner than their names and therefore yield faster inferences. Method Subjects and design. The same four subjects from Experiment 7 were used in the present experiment ; there was about a 6-week delay between experiments. All four subjects were tested in the ++ condition. Two subjects received 3 days o{ name stimuli and then 3 days of shape stimuli and the other two subjects received the reverse order. There were two practice blocks and eight experimental blocks per day. One subject in each order had previously been in the condition. The procedure was otherwise identical to that in Experiment 7. Materials. We constructed a new population of shape stimuli, in which the -western states (Alaska and Arizona) were approximately twice as large as the eastern states (New York and Florida), and the northern states (Alaska and New York) had a solid border, while the southern states (Arizona and Florida) had a dotted border. Thus, size and border were physical dimensions that were redundant with .geographic location.

Results and Discussion The error rates, averaged across subjects, were 2.7% and 3.4% for shapes and names, respectively. The data are shown in Figure

3; it can be seen that reaction time to name stimuli hardly improves over days, while reaction time to shape stimuli decreases. A mean score was obtained for each of the last eight blocks of trials on the third day for each stimulus type, with errors omitted. The data were analyzed in a mixed design, with blocks (1-8) and stimulus type (shapes or names) the within-subjects variables and order (shapes first or names first) the between-subjects variable. Both the main effect of stimulus type, F(\, 2) = 29.71, and the Stimulus Type X Order interaction, ^(l, 2) = 12.21, Af5e = 948.64, were reliable. The main effect shows that shapes were reliably faster than names (555 versus 596 msec, respectively). The interaction is due to the fact that the two subjects who received shape stimuli first did not show any further improvement at all when switched to name stimuli (their mean reaction times for Day 3 shapes and Day 6 names were 582 and 597 msec), while the subjects who received name stimuli first were able to benefit from the perceptual features after the switch (their means for Day 3 names and Day 6 shapes were 595 and 529 msec).

188

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

The results of Experiments 7 and 8 imply that pictures can function like names whenever they are representations of unique (i.e., proper noun) objects. For present purposes, however, the important point is that redundant physical attributes which may be used as an aid in discriminating the stimuli from one another, in fact, produce faster inferences. General Discussion In these experiments, we have used a speeded-inference task to contrast memorymatch models of encoding with a levels-ofprocessing analysis. We chose this task because statements about the depth of analysis required for performance are ambiguous when we are not sure of the basis for the response (e.g., the "CC < Aa" difference for a name match is about the same magnitude as the "color < size" difference for a physical match [Posner, 1969]), or when we cannot be sure that effects we are attributing to different depths may in fact result from an interaction between encoding and other operations (e.g., Thorndike-Lorge word frequency has a different effect on picture-word same judgments than it does on naming the pictures [Wingfield, 1968]). With a speeded-inference task, we can be certain of the outcome of the stimulus encoding process because the subject must explicitly name the basis of his response. Furthermore, the task has a built-in check for the independence of encoding and other processing operations. In general, the data of these experiments speak against any theory which assumes that pictures and words have differential access to memory representations (e.g., imaginal versus verbal) which comprise different forms of information (e.g., spatial versus generic). Picture-word differences in reaction time, where they exist in a speededinference task, are traceable to differences in stimulus discriminability and not to differences in the kind of information or code the stimulus contacts most readily. The clearest and probably most important implication in these data pertains to the relationship between reaction time and depth of processing necessary to abstract the stimulus

characteristics required by the task. Once a stimulus has been identified, further informational analysis proceeds along whichever dimensions are relevant to the performance requirement. Information about physical features is generally available sooner than semantic or conceptual information, regardless of the modality of the stimulus. While it may be that different memory representations do exist for physical and conceptual information, it seems unlikely that these codes are differentially activated by pictures and words for purposes of this task. Rather, it would appear as though the stimuli themselves, regardless of their modality, embody different levels of information which are encoded and then used as required. Inherent in both pictures and words is information about the concept they represent, either as an instance (picture) or as a name (word). Our analysis implies that encoding a stimulus first entails discriminating it from among the set of possible alternatives and then, as the task requires, further analyzing it along whatever dimensions (physical or semantic) are necessary. We must again stress that we do not wish our use of the term identification to carry with it the implication that the stimulus is named or matched to a memory representation; rather, identification is tacit. Variables which may aid the identification of a particular stimulus do so by allowing for better discrimination among all the stimuli in a set. For example, in Experiments 1 and 2, we found a pictureword difference in favor of pictures when stimuli had to be processed to the same semantic level. The picture-word difference in these experiments cannot be a depth effect because the response requirements (the solution-relevant characteristics) are the same for both groups. It is instead due to the fact that under most circumstances, pictures are more readily distinguishable from one another than are words, giving them a head start in the encoding process. Experiment 8 lends further support to this idea, since we were able to produce a reaction time difference by adding physical features to the shapes of states that were redundant with, but different from, : the conceptual information required for performance.

ENCODING PICTURES AND WORDS

189

TABLE 9 REACTION TIMES FOR ++ INFERENCES AS A FUNCTION OF THE DEPTH OF THE REQUIRED RESPONSES Experiment

8 (Day 3)

5 (Block 8) 4 (Block 8) Van Rijn (1973) 1 ahd-2 (Block 8) 7 (Day 1)

1 and 2 (Block 8) 4 (Block 8)

Stimulus materials

pictures (states with size and border dimensions) words (colored ink) pictures (nonsense creatures) geometric designs pictures (hippo, mouse, bus, car) pictures and words (states) words (hippo, mouse, bus, car) pictures (nonsense creatures)

In Table 9 we list, in order of increasing magnitude, the reaction times for the + + conditions of seven different experiments comprising a wide range of stimulus materials. While comparisons across experiments should be made with caution, the range of the reaction time difference between inference conditions is only 94 msec, which is, in the context of this task, truly remarkable. We may therefore appropriately consider the reaction time differences across stimulus materials to be "purely" due to encoding factors, and try to find some regularity within them which would explain the particular ordering obtained. Scanning down the "Stimulus materials" column gives us no particular clue to the ordering of the reaction times; pictures of objects, words, states, and geometric designs seem to be randomly interleaved with one another, and the same stimulus materials may be found at the beginning and the end of the list. It is only when we look at the "Response" column that the data begin to make sense. It is generally the case that responses or relevant values which correspond to physical properties of the stimuli produce snorter reaction times than do responses which reflect conceptual properties, regardless of whether the materials are pictures or words.

Response

Reaction time

( + +)

(X + +) - (X--1

NORTH, SOUTH, EAST, WEST

555

—•

RED, YELLOW

638 730

300 275

RED, YELLOW, SQUARE, TRIANGLE

761

273

ANIMAL, VEHICLE, LARGE, SMALL

811

262

NORTH, SOUTH, EAST, WEST

817

329

ANIMAL, VEHICLE, LARGE, SMALL

952

251

1,070

235

HORNS, TAIL, ROUND, SQUARE

ALIVE, EXTINCT

Experiments 4 and 5 explicitly examined the reaction-time-informational-depth trend reflected in Table 9, the former using pictorial materials and the latter using words. As we have seen, the effect, demonstrated within subjects, was extremely powerful. Identifiability (as previously denned) cannot be responsible for encoding differences because the stimuli were the same for all subjects within each experiment. Experiments 6 and 7 similarly demonstrated that the location of a state, which we consider to be a relatively conceptual property, is as derivable from its name as from its shape; this result is consistent with the present analysis and also with the notion that shapes of states may function at the generic (symbolic) level. While performance differences to various types of stimuli are often taken as evidence for memory representations which contain qualitatively different kinds of information, this interpretation results from viewing the recognition process as a match between current input and stored representation. We have shown that by changing the required depth of processing, we can create performance differences within both pictorial and verbal materials. Within a levels-of-processing framework, stimuli are coded along

190

ALINDA FRIEDMAN AND LYLE E. BOURNE, JR.

whichever dimensions are relevant for performance, and the stored representations need not be contacted in order for the results of the current coding process to be available for subsequent use. The notion of levels of information derives from the idea that under most circumstances, physical codes are available for further use prior to name or conceptual codes. The fact that a subject can tap into his coding process at a variety of different depths suggests that recognition occurs at the point during encoding that is most appropriate for a given context, REFERENCE NOTE 1. Paivio, A. Images, propositions, and knowledge (Research Bulletin No. 309). London, Canada: University of Western Ontario, 1974. REFERENCES Craik, F. I, M., & Lockhart, R. S. Levels of processing: A framework for memory research. Journal of Verbal Learning and Verbal Behavior, 1972, 11, 671-684. Daehler, M. W. Developmental and experimental factors associated with inferential behavior. Journal of Experimental Child Psychology, 1972, 13, 324-338. Huttenlocher, J. Some effects of negative instances on the formation of simple concepts. Psychological Reports, 1962, 11, 35-42. Moyer, R. S. Comparing objects in memory: Evidence suggesting an internal psychophysics. Perception & Psychophysics, 1973, 13, 180-184. Paivio, A. Imagery and verbal processes. New York: Holt, Rinehart & Winston, 1971.

Paivio, A. Perceptual comparisons through the mind's eye. Memory & Cognition, 197S, 3, 635647. Paivio, A., & Begg, I. Pictures and words in visual search. Memory & Cognition, 1974, 2, 515-521. Posner, M. Abstraction and the process of recognition. In G. H. Bower & J. T. Spence (Eds.), The psychology of learning and motivation (Vol. 3). New York: Academic Press, 1969. Posner, M. Coordination of internal codes. In W. G. Chase (Ed.), Visual information processing. New York: Academic Press, 1973. Posner, M., Boies, S., Eichelman, W., & Taylor, R. Retention of visual and name codes of single letters. Journal of Experimental Psychology Monograph, 1969, 79(1, Ft. 2). Posner, M., & Mitchell, R. F. Chronometric analysis of classification. Psychological Review, 1967, 74, 392-409. Rosch, E. Cognitive representations of semantic categories. Journal of Experimental Psychology: General, 1975, 104, 192-233. Rosch, E., Mervis, C. B., Gray, W., Johnson, D., & Boyes-Braem, P. Basic objects in natural categories. Journal of Experimental Psychology: General, in press. Sternberg, S. Memory-scanning: Mental processes revealed by reaction-time experiments. American Scientist, 1969, 57, 421-457. Turvey, M. T. Constructive theory, perceptual systems, and tacit knowledge. In W. B. Weimer & D. S. Palermo (Eds.), Cognition and the symbolic processes. New York: Wiley, 1974. Van Rijn, P. An information processing analysis of a two-trial inference task. Unpublished doctoral dissertation, University of Colorado, 1973. Wingfield, A. Effects of frequency on identification and naming of objects. American Journal of Psychology, 1968, 81, 226-234. (Received August 6, 1975)