2009-02-23 06:57:41 -07:00
|
|
|
#include <linux/kvm_host.h>
|
|
|
|
#include <linux/kvm.h>
|
|
|
|
#include <linux/hrtimer.h>
|
|
|
|
#include <asm/atomic.h>
|
|
|
|
#include "kvm_timer.h"
|
|
|
|
|
|
|
|
static int __kvm_timer_fn(struct kvm_vcpu *vcpu, struct kvm_timer *ktimer)
|
|
|
|
{
|
|
|
|
int restart_timer = 0;
|
|
|
|
wait_queue_head_t *q = &vcpu->wq;
|
|
|
|
|
2009-06-09 07:37:01 -06:00
|
|
|
/*
|
|
|
|
* There is a race window between reading and incrementing, but we do
|
|
|
|
* not care about potentially loosing timer events in the !reinject
|
2010-03-23 11:15:53 -06:00
|
|
|
* case anyway. Note: KVM_REQ_PENDING_TIMER is implicitly checked
|
|
|
|
* in vcpu_enter_guest.
|
2009-06-09 07:37:01 -06:00
|
|
|
*/
|
|
|
|
if (ktimer->reinject || !atomic_read(&ktimer->pending)) {
|
2009-06-09 07:37:03 -06:00
|
|
|
atomic_inc(&ktimer->pending);
|
2009-06-09 07:37:01 -06:00
|
|
|
/* FIXME: this code should not know anything about vcpus */
|
2009-06-09 07:37:03 -06:00
|
|
|
set_bit(KVM_REQ_PENDING_TIMER, &vcpu->requests);
|
2009-06-09 07:37:01 -06:00
|
|
|
}
|
2009-02-23 06:57:41 -07:00
|
|
|
|
|
|
|
if (waitqueue_active(q))
|
|
|
|
wake_up_interruptible(q);
|
|
|
|
|
|
|
|
if (ktimer->t_ops->is_periodic(ktimer)) {
|
|
|
|
hrtimer_add_expires_ns(&ktimer->timer, ktimer->period);
|
|
|
|
restart_timer = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return restart_timer;
|
|
|
|
}
|
|
|
|
|
|
|
|
enum hrtimer_restart kvm_timer_fn(struct hrtimer *data)
|
|
|
|
{
|
|
|
|
int restart_timer;
|
|
|
|
struct kvm_vcpu *vcpu;
|
|
|
|
struct kvm_timer *ktimer = container_of(data, struct kvm_timer, timer);
|
|
|
|
|
2009-06-09 06:56:27 -06:00
|
|
|
vcpu = ktimer->vcpu;
|
2009-02-23 06:57:41 -07:00
|
|
|
if (!vcpu)
|
|
|
|
return HRTIMER_NORESTART;
|
|
|
|
|
|
|
|
restart_timer = __kvm_timer_fn(vcpu, ktimer);
|
|
|
|
if (restart_timer)
|
|
|
|
return HRTIMER_RESTART;
|
|
|
|
else
|
|
|
|
return HRTIMER_NORESTART;
|
|
|
|
}
|
|
|
|
|