Caliban ignored him, and continued his survey of the room. There was a portable refresher unit in Beddle’s side of the room, and a large supply of bottled water and survival rations stacked up on the opposite side of the room from the refresher. A primitive cot, with one blanket and one pillow, stood in the center of the cell.
And Caliban understood. The torpedo-shaped device was, of course, the burrow bomb. It was hooked up to the photosensors. If Beddle tried to step across the sensor barrier, the bomb would go up-or at least Prospero had convinced him that it would. It came to much the same thing.
But Caliban understood more than that. A robot may not injure a human being. That was the New First Law, in its entirety. And, at least by the most parsimonious and niggardly of interpretations, Prospero had not in literal fact harmed Beddle. No doubt he had carried some utterly safe anesthetic with him when he had hidden himself aboard Beddle’ s aircar. He had seen to it that the unconscious Beddle had plenty of air for his ride across the lakebed in the cargo roller. And he had provided Beddle with ample food and water, adequate sanitation facilities, serviceable clothes, and a decent bed. He had done the man no harm at all, at least in any literal, physical sense.
And if Beddle elected to stay where he was, he would not come to any harm at Prospero’s hand. And if he crossed the optical sensor barrier, it would be Beddle’s action-not Prospero’s-that would set off the bomb and destroy him. Beddle would kill himself with the bomb he had meant to use to kill a city full of New Law robots.
And Prospero would not be forced to interfere. The second clause of the original First Law required a robot to take action to prevent harm. A Three-Law robot could not stand idly by if Beddle endangered himself. But not so the New Law robots. Prospero could, through inaction, allow a human to come to harm.
And when the comet struck then Beddle would die, yes, but not through any action of Prospero’s. It would be the actions of others-of Davlo Lentrall, of Alvar Kresh, of all the engineers and designers and pilots who moved the comet-that killed him. It would not be Prospero.
Prospero had found a loophole in the New First Law. He had found a way to kill without killing. All it required was as miserly-and as vicious-a parsing of the New First Law as Caliban could imagine.
And it also required Prospero to be half mad, at least. The leader of the New Law robots turned to face Caliban, and it was instantly obvious that Prospero could meet that requirement without the slightest difficulty. His orange eyes glowed with too brilliant a fire. The fingers of his left hand were twitching spasmodically. Dealing with his parsimonious interpretation New First Law had clearly imposed a tremendous amount of stress. And clearly, Prospero had cracked under the pressure. “Caliban!” he cried out, a wild pleasure in his voice. “I knew it would be you. I knew if anyone figured it out, it would be you.”
“Prospero, you are insane,” Caliban said. “Stop this. Stop this now, and let us all depart.”
“How did you figure it out?” Prospero asked, completely ignoring what Caliban had said. He turned more fully toward Caliban, moving a bit too quickly, and nearly overbalanced himself. “What was the clue that led you here?”
“Norlan Fiyle said that whoever killed the robots at the aircar hated Three Law robots. You have always held them in contempt.”
“Willing slaves,” Prospero said. “Collaborators in their own oppression. They don’t matter.”
“And what of Lancon-03 and the other New Law robots that lie dead in the halls of Valhalla?”
“Unfortunate, but necessary. They would have interfered. They would have stopped me. I had to choose the greatest good for the greatest number. Now they cannot stop me.” Prospero’s gaze shifted to the desk behind him. There was a blaster on it.
Caliban ignored the implied threat. “I can stop you,” he said. “I will.”
“No,” said Prospero. “No you can’t. You won’t.”
“I have no choice,” said Caliban. “If I can deduce the truth, so will others. The moment the humans realize that a New Law robot engineered the death of a human being, the New Law robots will be exterminated.”
“I have not engineered his death!” Prospero protested in a voice that suddenly turned shrill. “I have not harmed a human being. I…I merely offered choices to others.”