We’re pretty good at remembering objects in a complex scene. We can even remember those objects when we move to a different location. However, the research so far has found that memory for the original view is a little better than memory when we’ve moved to a different location. Much of that research, however, has focused on relatively complex movements: Viewers are asked to remember an array of objects viewed from one side of a room, then are transported to a different part of the room and asked to decide whether the objects are arranged in the same pattern (actually, they’re sitting at a computer watching static images, but the camera has moved).
This sort of motion actually encompasses two separate motions: translation and rotation. For example, in this crude figure representing a room, a viewer moving from the bottom to the right-hand side of the room would have to not only walk to a different part of the room, but also rotate his body (or at least his head) to see the objects in the room.
So what about simpler forms of motion? Do translating and rotating alone affect memory for the arrangement of objects in a scene? Rotating is difficult to assess independently, since if you rotate your body too much, you’re no longer viewing the scene. But David Waller was able to consider two different types of translation in scenes: backwards and forwards, and side-to-side.
First, Waller trained viewers to recall the arrangement of five toys in a photograph. Then the photograph was systematically altered by either moving the camera towards or away from the toys, or side-to-side:
In every case, all the toys were still visible, they were just closer to the viewer or farther away, or skewed to one side or another. He created similar sets of photos for 10 different arrangements of the toys. Viewers were then shown all the photos and asked which ones were in the correct, original pattern. Here are the results:
The graph shows accuracy when the scene matched the scene viewers had been trained on. Even for translations alone with no rotation, viewers were significantly more accurate when there was no shift in perspective. But notice that there’s one exception: When the viewpoint moved back from the trained viewpoint, there was no drop-off in accuracy. Waller believes this result may be due to boundary extension: We misremember the boundaries of a scene as wider than they really are, so when the camera moves back, we think we’re seeing the same scene. There were similar results for reaction times.
But maybe the results for side-to-side displacement are due to something else: maybe it’s just harder for us to remember the locations of objects that aren’t centered in a scene. So in another experiment, Waller trained some viewers on centered scenes, and other viewers on scenes where the objects were clustered to the left or to the right. So there are three possible positions for the objects — center, right, or left of the screen. But if you were trained on objects clustered to the left or the right, then during testing, you would see objects either at the center of the screen or on the opposite side of the screen from where you were trained. Here are those results:
The accuracy rates are lower than for the previous experiment because these images were flashed only briefly — just 250 milliseconds — but when viewers learned the centered scene, the results matched previous results. Accuracy was higher for the scene they had learned than for displaced scenes. Surprisingly, however, the results were different when viewers learned the scenes that had been displaced to one side: They were just as accurate identifying scenes that were displaced relative to training as they were on the scenes they had actually learned. Only when the scenes were shifted to the opposite side of the screen did accuracy decrease significantly. This suggests that there is a bias towards recalling parts of a scene that are closest to the center.
This isn’t just due to a tendency to focus on the center of the screen. Before each scene was showed to viewers, a cross appeared to direct their attention to the appropriate part of the computer screen.
Waller says this suggests that our memory for scenes may not include a true three-dimensional representation of the objects in the scene. Instead, we may simply recall “snapshots” taken from the perspective at which we were trained. When items are centered in the snapshot, or in another snapshot we’re asked to compare, it’s easier for us to recall whether they are the same. If we had a true three-dimensional representation of the scene, it’s difficult to see how Waller would find the results he did.
Waller, D. (2006). Egocentric and nonegocentric coding in memory for spatial layout: Evidence from scene recognition. Memory & Cognition, 34(3), 491-504.