From: Marcelo Tosatti on 13 Jul 2010 17:00 On Mon, Jul 12, 2010 at 04:25:29PM -1000, Zachary Amsden wrote: > Make the match of TSC find TSC writes that are close to each other > instead of perfectly identical; this allows the compensator to also > work in migration / suspend scenarios. > > Signed-off-by: Zachary Amsden <zamsden(a)redhat.com> > --- > arch/x86/kvm/x86.c | 14 ++++++++++---- > 1 files changed, 10 insertions(+), 4 deletions(-) > > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index 79c4608..51d3f3e 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -926,21 +926,27 @@ void guest_write_tsc(struct kvm_vcpu *vcpu, u64 data) > struct kvm *kvm = vcpu->kvm; > u64 offset, ns, elapsed; > struct timespec ts; > + s64 sdiff; > > spin_lock(&kvm->arch.tsc_write_lock); > offset = data - native_read_tsc(); > ns = get_kernel_ns(); > elapsed = ns - kvm->arch.last_tsc_nsec; > + sdiff = data - kvm->arch.last_tsc_write; > + if (sdiff < 0) > + sdiff = -sdiff; > > /* > - * Special case: identical write to TSC within 5 seconds of > + * Special case: close write to TSC within 5 seconds of > * another CPU is interpreted as an attempt to synchronize > - * (the 5 seconds is to accomodate host load / swapping). > + * The 5 seconds is to accomodate host load / swapping as > + * well as any reset of TSC during the boot process. > * > * In that case, for a reliable TSC, we can match TSC offsets, > - * or make a best guest using kernel_ns value. > + * or make a best guest using elapsed value. > */ > - if (data == kvm->arch.last_tsc_write && elapsed < 5ULL * NSEC_PER_SEC) { > + if (sdiff < nsec_to_cycles(5ULL * NSEC_PER_SEC) && > + elapsed < 5ULL * NSEC_PER_SEC) { > if (!check_tsc_unstable()) { > offset = kvm->arch.last_tsc_offset; > pr_debug("kvm: matched tsc offset for %llu\n", data); What prevents a vcpu from seeing its TSC go backwards, in case the first write in the 5 second window is smaller than the victim vcpu's last visible TSC value ? -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Zachary Amsden on 13 Jul 2010 17:20 On 07/13/2010 10:34 AM, Marcelo Tosatti wrote: > On Mon, Jul 12, 2010 at 04:25:29PM -1000, Zachary Amsden wrote: > >> Make the match of TSC find TSC writes that are close to each other >> instead of perfectly identical; this allows the compensator to also >> work in migration / suspend scenarios. >> >> Signed-off-by: Zachary Amsden<zamsden(a)redhat.com> >> --- >> arch/x86/kvm/x86.c | 14 ++++++++++---- >> 1 files changed, 10 insertions(+), 4 deletions(-) >> >> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c >> index 79c4608..51d3f3e 100644 >> --- a/arch/x86/kvm/x86.c >> +++ b/arch/x86/kvm/x86.c >> @@ -926,21 +926,27 @@ void guest_write_tsc(struct kvm_vcpu *vcpu, u64 data) >> struct kvm *kvm = vcpu->kvm; >> u64 offset, ns, elapsed; >> struct timespec ts; >> + s64 sdiff; >> >> spin_lock(&kvm->arch.tsc_write_lock); >> offset = data - native_read_tsc(); >> ns = get_kernel_ns(); >> elapsed = ns - kvm->arch.last_tsc_nsec; >> + sdiff = data - kvm->arch.last_tsc_write; >> + if (sdiff< 0) >> + sdiff = -sdiff; >> >> /* >> - * Special case: identical write to TSC within 5 seconds of >> + * Special case: close write to TSC within 5 seconds of >> * another CPU is interpreted as an attempt to synchronize >> - * (the 5 seconds is to accomodate host load / swapping). >> + * The 5 seconds is to accomodate host load / swapping as >> + * well as any reset of TSC during the boot process. >> * >> * In that case, for a reliable TSC, we can match TSC offsets, >> - * or make a best guest using kernel_ns value. >> + * or make a best guest using elapsed value. >> */ >> - if (data == kvm->arch.last_tsc_write&& elapsed< 5ULL * NSEC_PER_SEC) { >> + if (sdiff< nsec_to_cycles(5ULL * NSEC_PER_SEC)&& >> + elapsed< 5ULL * NSEC_PER_SEC) { >> if (!check_tsc_unstable()) { >> offset = kvm->arch.last_tsc_offset; >> pr_debug("kvm: matched tsc offset for %llu\n", data); >> > What prevents a vcpu from seeing its TSC go backwards, in case the first > write in the 5 second window is smaller than the victim vcpu's last > visible TSC value ? > Nothing, unfortunately. However, the TSC would already have to be out of sync in order for the problem to occur. It can never happen in normal circumstances on a stable hardware TSC except in one case; migration. During the CPU state transfer phase of migration, however, all the VCPUs should already be stopped, so the maximum TSC that can be observed by any CPU is bounded. The problem, of course is that the TSC write will latch the first TSC to be written, which, if you stop in order, and start in order, will be the lowest TSC; so later VCPUs can observe a negative TSC drift (timing is additionally complicated by the VCPU teardown / create time). There are a couple solutions; some of which are ugly and some of which are very ugly. 1) Make some global state available about whether the VM is running or not, use the global TSC lock and record the last TSC when the VM is stopped. Return this TSC from any reads when the VM is stopped. I'm not sure the kernel model is equipped to do this properly; in theory, userspace could stop and start running VCPUs using the ioctls whenever it feels like and requires no protocol to do so. 2) Make userspace deal with it; when starting up a VM, read the VCPU state for all VCPUs in first, then take the maximum TSC and set all TSCs to this value before starting the VCPUs. I'm not sure the userspace model is equipped to do this properly, it could start running earlier CPUs before reading later CPU states... 3) Drop passthrough TSC altogether and switch to trap / emulate TSC. 4) Pray to a deity of your choice. Of course, none of these solutions work for a guest which deliberately runs with desynchronized TSCs, but we needn't really be concerned with that, no guest does it on purpose. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Zachary Amsden on 13 Jul 2010 17:50 On 07/13/2010 11:42 AM, David S. Ahern wrote: > > On 07/13/10 15:15, Zachary Amsden wrote: > > >>> What prevents a vcpu from seeing its TSC go backwards, in case the first >>> write in the 5 second window is smaller than the victim vcpu's last >>> visible TSC value ? >>> >>> >> Nothing, unfortunately. However, the TSC would already have to be out >> of sync in order for the problem to occur. It can never happen in >> normal circumstances on a stable hardware TSC except in one case; >> migration. During the CPU state transfer phase of migration, however, >> > What about across processor sockets? Aren't CPUs brought up at different > points such that their TSCs start at different times? > Yes, that's called an unsynchronized TSC. In that case, the compensation does the best it can based on time since the first TSC write, but it will never be exact. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: David S. Ahern on 13 Jul 2010 18:00 On 07/13/10 15:15, Zachary Amsden wrote: >> What prevents a vcpu from seeing its TSC go backwards, in case the first >> write in the 5 second window is smaller than the victim vcpu's last >> visible TSC value ? >> > > Nothing, unfortunately. However, the TSC would already have to be out > of sync in order for the problem to occur. It can never happen in > normal circumstances on a stable hardware TSC except in one case; > migration. During the CPU state transfer phase of migration, however, What about across processor sockets? Aren't CPUs brought up at different points such that their TSCs start at different times? David -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Zachary Amsden on 13 Jul 2010 19:40 On 07/13/2010 11:42 AM, David S. Ahern wrote: > > On 07/13/10 15:15, Zachary Amsden wrote: > > >>> What prevents a vcpu from seeing its TSC go backwards, in case the first >>> write in the 5 second window is smaller than the victim vcpu's last >>> visible TSC value ? >>> >>> >> Nothing, unfortunately. However, the TSC would already have to be out >> of sync in order for the problem to occur. It can never happen in >> normal circumstances on a stable hardware TSC except in one case; >> migration. During the CPU state transfer phase of migration, however, >> > What about across processor sockets? Aren't CPUs brought up at different > points such that their TSCs start at different times? > It depends on the platform. But it doesn't matter. The definition we use is different start TSCs == out of sync. Some systems have synchronized TSCs, some do not. See patch 18/18 - "Timekeeping documentation" for details. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
|
Next
|
Last
Pages: 1 2 Prev: x86, memblock: Use memblock_debug to control debug message print out Next: Build log scripts |