Re: VMI Interface Proposal Documentation for I386, Part 5

From: Zachary Amsden
Date: Tue Mar 14 2006 - 03:23:29 EST


Zwane Mwaikambo wrote:
On Mon, 13 Mar 2006, Zachary Amsden wrote:

PROCESSOR STATE CALLS

This set of calls controls the online status of the processor. It
include interrupt control, reboot, halt, and shutdown functionality.
Future expansions may include deep sleep and hotplug CPU capabilities.

VMI_DisableInterrupts

VMICALL void VMI_DisableInterrupts(void);

Disable maskable interrupts on the processor.

Inputs: None
Outputs: None
Clobbers: Flags only
Segments: As this is both performance critical and likely to
be called from low level interrupt code, this call does not
require flat DS/ES segments, but uses the stack segment for
data access. Therefore only CS/SS must be well defined.

VMI_EnableInterrupts

VMICALL void VMI_EnableInterrupts(void);

Enable maskable interrupts on the processor. Note that the
current implementation always will deliver any pending interrupts
on a call which enables interrupts, for compatibility with kernel
code which expects this behavior. Whether this should be required
is open for debate.

Mind if i push this debate slightly forward? If we were to move the dispatch of pending interrupts elsewhere, where would that be? In particular, for a device which won't issue any more interrupts until it's previous interrupt is serviced. Perhaps injection at arbitrary points during runtime when interrupts are enabled?

Thanks for the response.

This is exactly what I was hoping for - discussion. Think about this from the hypervisor perspective - if the guest enables interrupts, and you have something pending to deliver, for correctness, you have to deliver it, right now. But does the kernel truly require that interrupt deliver immediately - in most cases, no. In particular, on the fast path for system calls, one of the first instructions executed is "STI." Do you really want to take interrupts there? No, but you have to let them come in. So you work around that fact by allowing them, even if it inconveniences you. In some cases, you have not yet set up even proper kernel segments to access data.

It could be possible to change the semantics of the interrupt masking interface in Linux, such that enable_interrupts() did just that - but did not yet deliver pending IRQs. As did restore_interrupt_mask(). This would require inspection of many drivers to ensure that they don't rely on those actions causing immediate interrupt delivery. And if they did, they would require a call, say, deliver_pending_irqs() to accomplish that.

Is this a nice interface for Linux? Probably not. In fact, requiring source inspection of all drivers just for this would be a gargantuan task, as well as being difficult to maintain. Perhaps, it may have some benefit - one ideology is that drivers should not in general require the ability to enable and receive interrupts immediately. Otherwise, they are dependent on hardware responses to continue operation, which means they are probably not fault tolerant / recoverable. But many drivers have been written this way.

The motivation here is entirely selfish. Emulating the CPU by unquestioning delivery of interrupts is a fine course of action - but it does impose a slight overhead. You first have to determine if there are any interrupts / callbacks / upcalls to be serviced. This is not something you can do in one instruction, and moreover, you may have to deal with race conditions in determining whether or not any actions are pending. So there is a measurable benefit, when running in a virtual machine, to separate the required delivery or interrupts with the enabling of them.

That is why I think it warrants discussion on the principles, although I am not sure that it is practical.

Zach
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/