When we read a Jane Austen novel, what we look at is just a myriad of black smudges arranged neatly in lines on a set of white rectangles, and yet what we feel we are “seeing” (and should I use the quotation marks or not?) is a mansion in the English countryside, a team of horses pulling a carriage down a country lane, an elegantly clad lady and gentleman sitting side by side in the carriage exchanging pleasantries when they espy a poor old woman emerging from her humble cottage along the roadside… We are so taken in by what we “see” that in some important and serious sense we don’t notice the room we are sitting in, the trees visible through its window, nor even the black smudges speckled all over the white rectangles in our hands (even though, paradoxically, we are depending on those smudges to bring us the visual images I just described). If you don’t believe me, consider what you have just been doing in the last thirty seconds: processing black smudges speckled on white rectangles and yet “seeing” someone reading a Jane Austen novel in an armchair in a living room, and in addition, seeing the mansion, the country road, the carriage, the elegant couple, and the old woman… Black curlicues on a white background, when suitably arranged, transport us in milliseconds to arbitrarily distant, long-gone, or even never-existent venues and epochs.
The point of all of this is to insist on the idea that we
Where Am I?
As I was driving a few days ago, I pulled up alongside a jogger waiting at a red light. She was trotting in place, and then the light changed and she crossed the street and disappeared. For a brief moment, I was “in her shoes”. I had never seen her before and probably will never see her again, but I have been there many a time. I had lived that experience in my own way, and even though I know virtually nothing about her, I have shared that experience of hers. To be sure, I was not seeing it through her eyes. But let’s briefly jump once again into the realm of slightly silly technological extravagance.
Suppose everyone wore a tiny TV camera on the bridge of their nose, and that everyone had glasses that could be tuned to receive the signals from any selected TV camera on earth. If there were a way of specifying a person by their GPS coordinates (and that certainly doesn’t seem far-fetched), then all I would have to do is set my glasses to receive the signals from that jogger’s nose-mounted TV camera, and presto! — I would suddenly be seeing the world from her perspective. When I was sitting in my car and the traffic light changed and she took off and disappeared, I could have ridden along and seen just where she was going, could have heard the birds chirping as she jogged through a woodsy lane, and so forth. And at any point I could switch channels and go see the world through the nose-camera of my daughter Monica or my son Danny, or anyone else I wished. So where am I? “Still just where you are!” chirps common sense. But that’s too simplistic, too ambiguous.
What determines “where I am”? If we once again postulate the idea of obtaining nutrition by carrying out certain remote actions, and if we add back the ability to control distant motion by means of a joystick or even by certain brain events, then things really start to shimmer in uncertainty. For surely a mobile robot is not where the radio-connected computer that is controlling it happens to be sitting. A robot might be strolling about on the moon while its computerized guidance system was in some earthbound laboratory. Or a self-driving car like Stanley could be crossing the Nevada desert, and its computer control system might be on board or might be located in a lab in California, connected by radio. But would we even care where the computer was? Why should we care where it is located?
A robot, we feel, is where its
Varying Degrees of Being Another