OSDN Git Service

x86/kvm: Always inline vmload() / vmsave()
authorPeter Zijlstra <peterz@infradead.org>
Thu, 24 Jun 2021 09:41:04 +0000 (11:41 +0200)
committerPeter Zijlstra <peterz@infradead.org>
Wed, 15 Sep 2021 13:51:45 +0000 (15:51 +0200)
vmlinux.o: warning: objtool: svm_vcpu_enter_exit()+0xea: call to vmload() leaves .noinstr.text section
vmlinux.o: warning: objtool: svm_vcpu_enter_exit()+0x133: call to vmsave() leaves .noinstr.text section

Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Link: https://lore.kernel.org/r/20210624095147.942250748@infradead.org
arch/x86/kvm/svm/svm_ops.h

index 22e2b01..9430d64 100644 (file)
@@ -56,12 +56,12 @@ static inline void invlpga(unsigned long addr, u32 asid)
  * VMSAVE, VMLOAD, etc... is still controlled by the effective address size,
  * hence 'unsigned long' instead of 'hpa_t'.
  */
-static inline void vmsave(unsigned long pa)
+static __always_inline void vmsave(unsigned long pa)
 {
        svm_asm1(vmsave, "a" (pa), "memory");
 }
 
-static inline void vmload(unsigned long pa)
+static __always_inline void vmload(unsigned long pa)
 {
        svm_asm1(vmload, "a" (pa), "memory");
 }