06a1338
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
06a1338
index 094b5d9..64a4b03 100644
06a1338
--- a/arch/x86/kvm/x86.c
06a1338
+++ b/arch/x86/kvm/x86.c
06a1338
@@ -1194,20 +1194,37 @@ void kvm_write_tsc(struct kvm_vcpu *vcpu, struct msr_data *msr)
06a1338
 	elapsed = ns - kvm->arch.last_tsc_nsec;
06a1338
 
06a1338
 	if (vcpu->arch.virtual_tsc_khz) {
06a1338
+		int faulted = 0;
06a1338
+
06a1338
 		/* n.b - signed multiplication and division required */
06a1338
 		usdiff = data - kvm->arch.last_tsc_write;
06a1338
 #ifdef CONFIG_X86_64
06a1338
 		usdiff = (usdiff * 1000) / vcpu->arch.virtual_tsc_khz;
06a1338
 #else
06a1338
 		/* do_div() only does unsigned */
06a1338
-		asm("idivl %2; xor %%edx, %%edx"
06a1338
-		: "=A"(usdiff)
06a1338
-		: "A"(usdiff * 1000), "rm"(vcpu->arch.virtual_tsc_khz));
06a1338
+		asm("1: idivl %[divisor]\n"
06a1338
+		    "2: xor %%edx, %%edx\n"
06a1338
+		    "   movl $0, %[faulted]\n"
06a1338
+		    "3:\n"
06a1338
+		    ".section .fixup,\"ax\"\n"
06a1338
+		    "4: movl $1, %[faulted]\n"
06a1338
+		    "   jmp  3b\n"
06a1338
+		    ".previous\n"
06a1338
+
06a1338
+		_ASM_EXTABLE(1b, 4b)
06a1338
+
06a1338
+		: "=A"(usdiff), [faulted] "=r" (faulted)
06a1338
+		: "A"(usdiff * 1000), [divisor] "rm"(vcpu->arch.virtual_tsc_khz));
06a1338
+
06a1338
 #endif
06a1338
 		do_div(elapsed, 1000);
06a1338
 		usdiff -= elapsed;
06a1338
 		if (usdiff < 0)
06a1338
 			usdiff = -usdiff;
06a1338
+
06a1338
+		/* idivl overflow => difference is larger than USEC_PER_SEC */
06a1338
+		if (faulted)
06a1338
+			usdiff = USEC_PER_SEC;
06a1338
 	} else
06a1338
 		usdiff = USEC_PER_SEC; /* disable TSC match window below */
06a1338