diff mbox series

[1/7] KVM: s390: clean up redundant 'kvm_run' parameters

Message ID 20200419075106.16248-2-tianjia.zhang@linux.alibaba.com (mailing list archive)
State New, archived
Headers show
Series clean up redundant 'kvm_run' parameters | expand

Commit Message

tianjia.zhang April 19, 2020, 7:51 a.m. UTC
In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
structure. Earlier than historical reasons, many kvm-related function
parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
This patch does a unified cleanup of these remaining redundant parameters.

Signed-off-by: Tianjia Zhang <tianjia.zhang@linux.alibaba.com>
---
 arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
 1 file changed, 67 insertions(+), 60 deletions(-)

Comments

Christian Borntraeger April 20, 2020, 7:07 a.m. UTC | #1
On 19.04.20 09:51, Tianjia Zhang wrote:
> In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
> structure. Earlier than historical reasons, many kvm-related function
> parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
> This patch does a unified cleanup of these remaining redundant parameters.
> 
> Signed-off-by: Tianjia Zhang <tianjia.zhang@linux.alibaba.com>
> ---
>  arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
>  1 file changed, 67 insertions(+), 60 deletions(-)
> 
> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
> index 443af3ead739..cf420d013ba3 100644
> --- a/arch/s390/kvm/kvm-s390.c
> +++ b/arch/s390/kvm/kvm-s390.c
> @@ -4173,24 +4173,25 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
>  	return rc;
>  }
>  
> -static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
>  {
> +	struct kvm_run *run = vcpu->run;

Please use kvm_run as variable name. This makes all of the changes below go away.


>  	struct runtime_instr_cb *riccb;
>  	struct gs_cb *gscb;
>  
> -	riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
> -	gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
> -	vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
> -	vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> -		vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
> -		vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
> -		vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
> -	}
> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
> -		vcpu->arch.pfault_token = kvm_run->s.regs.pft;
> -		vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
> -		vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
> +	riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
> +	gscb = (struct gs_cb *) &run->s.regs.gscb;
> +	vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
> +	vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
> +	if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> +		vcpu->arch.sie_block->todpr = run->s.regs.todpr;
> +		vcpu->arch.sie_block->pp = run->s.regs.pp;
> +		vcpu->arch.sie_block->gbea = run->s.regs.gbea;
> +	}
> +	if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
> +		vcpu->arch.pfault_token = run->s.regs.pft;
> +		vcpu->arch.pfault_select = run->s.regs.pfs;
> +		vcpu->arch.pfault_compare = run->s.regs.pfc;
>  		if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
>  			kvm_clear_async_pf_completion_queue(vcpu);
>  	}
> @@ -4198,7 +4199,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  	 * If userspace sets the riccb (e.g. after migration) to a valid state,
>  	 * we should enable RI here instead of doing the lazy enablement.
>  	 */
> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
> +	if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>  	    test_kvm_facility(vcpu->kvm, 64) &&
>  	    riccb->v &&
>  	    !(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
> @@ -4209,7 +4210,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  	 * If userspace sets the gscb (e.g. after migration) to non-zero,
>  	 * we should enable GS here instead of doing the lazy enablement.
>  	 */
> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
> +	if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>  	    test_kvm_facility(vcpu->kvm, 133) &&
>  	    gscb->gssm &&
>  	    !vcpu->arch.gs_enabled) {
> @@ -4218,10 +4219,10 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  		vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
>  		vcpu->arch.gs_enabled = 1;
>  	}
> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
> +	if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>  	    test_kvm_facility(vcpu->kvm, 82)) {
>  		vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
> -		vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
> +		vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
>  	}
>  	if (MACHINE_HAS_GS) {
>  		preempt_disable();
> @@ -4232,45 +4233,47 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  		}
>  		if (vcpu->arch.gs_enabled) {
>  			current->thread.gs_cb = (struct gs_cb *)
> -						&vcpu->run->s.regs.gscb;
> +						&run->s.regs.gscb;
>  			restore_gs_cb(current->thread.gs_cb);
>  		}
>  		preempt_enable();
>  	}
> -	/* SIE will load etoken directly from SDNX and therefore kvm_run */
> +	/* SIE will load etoken directly from SDNX and therefore run */
>  }
>  
> -static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void sync_regs(struct kvm_vcpu *vcpu)
>  {
> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
> -		kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
> -		memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
> +	struct kvm_run *run = vcpu->run;
> +
> +	if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
> +		kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
> +	if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
> +		memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
>  		/* some control register changes require a tlb flush */
>  		kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
>  	}
> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> -		kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
> -		vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
> +	if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> +		kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
> +		vcpu->arch.sie_block->ckc = run->s.regs.ckc;
>  	}
>  	save_access_regs(vcpu->arch.host_acrs);
> -	restore_access_regs(vcpu->run->s.regs.acrs);
> +	restore_access_regs(run->s.regs.acrs);
>  	/* save host (userspace) fprs/vrs */
>  	save_fpu_regs();
>  	vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
>  	vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
>  	if (MACHINE_HAS_VX)
> -		current->thread.fpu.regs = vcpu->run->s.regs.vrs;
> +		current->thread.fpu.regs = run->s.regs.vrs;
>  	else
> -		current->thread.fpu.regs = vcpu->run->s.regs.fprs;
> -	current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
> +		current->thread.fpu.regs = run->s.regs.fprs;
> +	current->thread.fpu.fpc = run->s.regs.fpc;
>  	if (test_fp_ctl(current->thread.fpu.fpc))
>  		/* User space provided an invalid FPC, let's clear it */
>  		current->thread.fpu.fpc = 0;
>  
>  	/* Sync fmt2 only data */
>  	if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
> -		sync_regs_fmt2(vcpu, kvm_run);
> +		sync_regs_fmt2(vcpu);
>  	} else {
>  		/*
>  		 * In several places we have to modify our internal view to
> @@ -4282,19 +4285,21 @@ static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  		 * do only accept the condition code from userspace.
>  		 */
>  		vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
> -		vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
> +		vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
>  						   PSW_MASK_CC;
>  	}
>  
> -	kvm_run->kvm_dirty_regs = 0;
> +	run->kvm_dirty_regs = 0;
>  }
>  
> -static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void store_regs_fmt2(struct kvm_vcpu *vcpu)
>  {
> -	kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
> -	kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
> -	kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
> -	kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
> +	struct kvm_run *run = vcpu->run;
> +
> +	run->s.regs.todpr = vcpu->arch.sie_block->todpr;
> +	run->s.regs.pp = vcpu->arch.sie_block->pp;
> +	run->s.regs.gbea = vcpu->arch.sie_block->gbea;
> +	run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>  	if (MACHINE_HAS_GS) {
>  		__ctl_set_bit(2, 4);
>  		if (vcpu->arch.gs_enabled)
> @@ -4310,39 +4315,41 @@ static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  	/* SIE will save etoken directly into SDNX and therefore kvm_run */
>  }
>  
> -static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void store_regs(struct kvm_vcpu *vcpu)
>  {
> -	kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
> -	kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
> -	kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
> -	memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
> -	kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
> -	kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
> -	kvm_run->s.regs.pft = vcpu->arch.pfault_token;
> -	kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
> -	kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
> -	save_access_regs(vcpu->run->s.regs.acrs);
> +	struct kvm_run *run = vcpu->run;
> +
> +	run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
> +	run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
> +	run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
> +	memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
> +	run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
> +	run->s.regs.ckc = vcpu->arch.sie_block->ckc;
> +	run->s.regs.pft = vcpu->arch.pfault_token;
> +	run->s.regs.pfs = vcpu->arch.pfault_select;
> +	run->s.regs.pfc = vcpu->arch.pfault_compare;
> +	save_access_regs(run->s.regs.acrs);
>  	restore_access_regs(vcpu->arch.host_acrs);
>  	/* Save guest register state */
>  	save_fpu_regs();
> -	vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
> +	run->s.regs.fpc = current->thread.fpu.fpc;
>  	/* Restore will be done lazily at return */
>  	current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
>  	current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
>  	if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
> -		store_regs_fmt2(vcpu, kvm_run);
> +		store_regs_fmt2(vcpu);
>  }
>  
>  int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>  {
> -	struct kvm_run *kvm_run = vcpu->run;
> +	struct kvm_run *run = vcpu->run;
>  	int rc;
>  
> -	if (kvm_run->immediate_exit)
> +	if (run->immediate_exit)
>  		return -EINTR;
>  
> -	if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
> -	    kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
> +	if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
> +	    run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>  		return -EINVAL;
>  
>  	vcpu_load(vcpu);
> @@ -4368,14 +4375,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>  		goto out;
>  	}
>  
> -	sync_regs(vcpu, kvm_run);
> +	sync_regs(vcpu);
>  	enable_cpu_timer_accounting(vcpu);
>  
>  	might_fault();
>  	rc = __vcpu_run(vcpu);
>  
>  	if (signal_pending(current) && !rc) {
> -		kvm_run->exit_reason = KVM_EXIT_INTR;
> +		run->exit_reason = KVM_EXIT_INTR;
>  		rc = -EINTR;
>  	}
>  
> @@ -4390,7 +4397,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>  	}
>  
>  	disable_cpu_timer_accounting(vcpu);
> -	store_regs(vcpu, kvm_run);
> +	store_regs(vcpu);
>  
>  	kvm_sigset_deactivate(vcpu);
>  
>
tianjia.zhang April 20, 2020, 7:17 a.m. UTC | #2
On 2020/4/20 15:07, Christian Borntraeger wrote:
> 
> 
> On 19.04.20 09:51, Tianjia Zhang wrote:
>> In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
>> structure. Earlier than historical reasons, many kvm-related function
>> parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
>> This patch does a unified cleanup of these remaining redundant parameters.
>>
>> Signed-off-by: Tianjia Zhang <tianjia.zhang@linux.alibaba.com>
>> ---
>>   arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
>>   1 file changed, 67 insertions(+), 60 deletions(-)
>>
>> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
>> index 443af3ead739..cf420d013ba3 100644
>> --- a/arch/s390/kvm/kvm-s390.c
>> +++ b/arch/s390/kvm/kvm-s390.c
>> @@ -4173,24 +4173,25 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
>>   	return rc;
>>   }
>>   
>> -static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
>>   {
>> +	struct kvm_run *run = vcpu->run;
> 
> Please use kvm_run as variable name. This makes all of the changes below go away.
> 

It's OK, I will fix it in v2 patch.

Thanks,
Tianjia

> 
>>   	struct runtime_instr_cb *riccb;
>>   	struct gs_cb *gscb;
>>   
>> -	riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
>> -	gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
>> -	vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
>> -	vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> -		vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
>> -		vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
>> -		vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
>> -	}
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
>> -		vcpu->arch.pfault_token = kvm_run->s.regs.pft;
>> -		vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
>> -		vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
>> +	riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
>> +	gscb = (struct gs_cb *) &run->s.regs.gscb;
>> +	vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
>> +	vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
>> +	if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> +		vcpu->arch.sie_block->todpr = run->s.regs.todpr;
>> +		vcpu->arch.sie_block->pp = run->s.regs.pp;
>> +		vcpu->arch.sie_block->gbea = run->s.regs.gbea;
>> +	}
>> +	if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
>> +		vcpu->arch.pfault_token = run->s.regs.pft;
>> +		vcpu->arch.pfault_select = run->s.regs.pfs;
>> +		vcpu->arch.pfault_compare = run->s.regs.pfc;
>>   		if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
>>   			kvm_clear_async_pf_completion_queue(vcpu);
>>   	}
>> @@ -4198,7 +4199,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>   	 * If userspace sets the riccb (e.g. after migration) to a valid state,
>>   	 * we should enable RI here instead of doing the lazy enablement.
>>   	 */
>> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>> +	if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>>   	    test_kvm_facility(vcpu->kvm, 64) &&
>>   	    riccb->v &&
>>   	    !(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
>> @@ -4209,7 +4210,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>   	 * If userspace sets the gscb (e.g. after migration) to non-zero,
>>   	 * we should enable GS here instead of doing the lazy enablement.
>>   	 */
>> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>> +	if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>>   	    test_kvm_facility(vcpu->kvm, 133) &&
>>   	    gscb->gssm &&
>>   	    !vcpu->arch.gs_enabled) {
>> @@ -4218,10 +4219,10 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>   		vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
>>   		vcpu->arch.gs_enabled = 1;
>>   	}
>> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>> +	if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>>   	    test_kvm_facility(vcpu->kvm, 82)) {
>>   		vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
>> -		vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
>> +		vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
>>   	}
>>   	if (MACHINE_HAS_GS) {
>>   		preempt_disable();
>> @@ -4232,45 +4233,47 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>   		}
>>   		if (vcpu->arch.gs_enabled) {
>>   			current->thread.gs_cb = (struct gs_cb *)
>> -						&vcpu->run->s.regs.gscb;
>> +						&run->s.regs.gscb;
>>   			restore_gs_cb(current->thread.gs_cb);
>>   		}
>>   		preempt_enable();
>>   	}
>> -	/* SIE will load etoken directly from SDNX and therefore kvm_run */
>> +	/* SIE will load etoken directly from SDNX and therefore run */
>>   }
>>   
>> -static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void sync_regs(struct kvm_vcpu *vcpu)
>>   {
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
>> -		kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
>> -		memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
>> +	struct kvm_run *run = vcpu->run;
>> +
>> +	if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
>> +		kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
>> +	if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
>> +		memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
>>   		/* some control register changes require a tlb flush */
>>   		kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
>>   	}
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> -		kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
>> -		vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
>> +	if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> +		kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
>> +		vcpu->arch.sie_block->ckc = run->s.regs.ckc;
>>   	}
>>   	save_access_regs(vcpu->arch.host_acrs);
>> -	restore_access_regs(vcpu->run->s.regs.acrs);
>> +	restore_access_regs(run->s.regs.acrs);
>>   	/* save host (userspace) fprs/vrs */
>>   	save_fpu_regs();
>>   	vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
>>   	vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
>>   	if (MACHINE_HAS_VX)
>> -		current->thread.fpu.regs = vcpu->run->s.regs.vrs;
>> +		current->thread.fpu.regs = run->s.regs.vrs;
>>   	else
>> -		current->thread.fpu.regs = vcpu->run->s.regs.fprs;
>> -	current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
>> +		current->thread.fpu.regs = run->s.regs.fprs;
>> +	current->thread.fpu.fpc = run->s.regs.fpc;
>>   	if (test_fp_ctl(current->thread.fpu.fpc))
>>   		/* User space provided an invalid FPC, let's clear it */
>>   		current->thread.fpu.fpc = 0;
>>   
>>   	/* Sync fmt2 only data */
>>   	if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
>> -		sync_regs_fmt2(vcpu, kvm_run);
>> +		sync_regs_fmt2(vcpu);
>>   	} else {
>>   		/*
>>   		 * In several places we have to modify our internal view to
>> @@ -4282,19 +4285,21 @@ static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>   		 * do only accept the condition code from userspace.
>>   		 */
>>   		vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
>> -		vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
>> +		vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
>>   						   PSW_MASK_CC;
>>   	}
>>   
>> -	kvm_run->kvm_dirty_regs = 0;
>> +	run->kvm_dirty_regs = 0;
>>   }
>>   
>> -static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void store_regs_fmt2(struct kvm_vcpu *vcpu)
>>   {
>> -	kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
>> -	kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
>> -	kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
>> -	kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>> +	struct kvm_run *run = vcpu->run;
>> +
>> +	run->s.regs.todpr = vcpu->arch.sie_block->todpr;
>> +	run->s.regs.pp = vcpu->arch.sie_block->pp;
>> +	run->s.regs.gbea = vcpu->arch.sie_block->gbea;
>> +	run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>>   	if (MACHINE_HAS_GS) {
>>   		__ctl_set_bit(2, 4);
>>   		if (vcpu->arch.gs_enabled)
>> @@ -4310,39 +4315,41 @@ static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>   	/* SIE will save etoken directly into SDNX and therefore kvm_run */
>>   }
>>   
>> -static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void store_regs(struct kvm_vcpu *vcpu)
>>   {
>> -	kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
>> -	kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
>> -	kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
>> -	memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
>> -	kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
>> -	kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
>> -	kvm_run->s.regs.pft = vcpu->arch.pfault_token;
>> -	kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
>> -	kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
>> -	save_access_regs(vcpu->run->s.regs.acrs);
>> +	struct kvm_run *run = vcpu->run;
>> +
>> +	run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
>> +	run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
>> +	run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
>> +	memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
>> +	run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
>> +	run->s.regs.ckc = vcpu->arch.sie_block->ckc;
>> +	run->s.regs.pft = vcpu->arch.pfault_token;
>> +	run->s.regs.pfs = vcpu->arch.pfault_select;
>> +	run->s.regs.pfc = vcpu->arch.pfault_compare;
>> +	save_access_regs(run->s.regs.acrs);
>>   	restore_access_regs(vcpu->arch.host_acrs);
>>   	/* Save guest register state */
>>   	save_fpu_regs();
>> -	vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
>> +	run->s.regs.fpc = current->thread.fpu.fpc;
>>   	/* Restore will be done lazily at return */
>>   	current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
>>   	current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
>>   	if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
>> -		store_regs_fmt2(vcpu, kvm_run);
>> +		store_regs_fmt2(vcpu);
>>   }
>>   
>>   int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>>   {
>> -	struct kvm_run *kvm_run = vcpu->run;
>> +	struct kvm_run *run = vcpu->run;
>>   	int rc;
>>   
>> -	if (kvm_run->immediate_exit)
>> +	if (run->immediate_exit)
>>   		return -EINTR;
>>   
>> -	if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
>> -	    kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>> +	if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
>> +	    run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>>   		return -EINVAL;
>>   
>>   	vcpu_load(vcpu);
>> @@ -4368,14 +4375,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>>   		goto out;
>>   	}
>>   
>> -	sync_regs(vcpu, kvm_run);
>> +	sync_regs(vcpu);
>>   	enable_cpu_timer_accounting(vcpu);
>>   
>>   	might_fault();
>>   	rc = __vcpu_run(vcpu);
>>   
>>   	if (signal_pending(current) && !rc) {
>> -		kvm_run->exit_reason = KVM_EXIT_INTR;
>> +		run->exit_reason = KVM_EXIT_INTR;
>>   		rc = -EINTR;
>>   	}
>>   
>> @@ -4390,7 +4397,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>>   	}
>>   
>>   	disable_cpu_timer_accounting(vcpu);
>> -	store_regs(vcpu, kvm_run);
>> +	store_regs(vcpu);
>>   
>>   	kvm_sigset_deactivate(vcpu);
>>   
>>
Janosch Frank April 20, 2020, 8:25 a.m. UTC | #3
On 4/20/20 9:07 AM, Christian Borntraeger wrote:
> 
> 
> On 19.04.20 09:51, Tianjia Zhang wrote:
>> In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
>> structure. Earlier than historical reasons, many kvm-related function
>> parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
>> This patch does a unified cleanup of these remaining redundant parameters.
>>
>> Signed-off-by: Tianjia Zhang <tianjia.zhang@linux.alibaba.com>
>> ---
>>  arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
>>  1 file changed, 67 insertions(+), 60 deletions(-)
>>
>> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
>> index 443af3ead739..cf420d013ba3 100644
>> --- a/arch/s390/kvm/kvm-s390.c
>> +++ b/arch/s390/kvm/kvm-s390.c
>> @@ -4173,24 +4173,25 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
>>  	return rc;
>>  }
>>  
>> -static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
>>  {
>> +	struct kvm_run *run = vcpu->run;
> 
> Please use kvm_run as variable name. This makes all of the changes below go away.

Definitely

@Tianjia
This patch produces way to much churn to be picked up as is.
I do not want patches that touch so much for small cosmetic changes.

> 
> 
>>  	struct runtime_instr_cb *riccb;
>>  	struct gs_cb *gscb;
>>  
>> -	riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
>> -	gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
>> -	vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
>> -	vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> -		vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
>> -		vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
>> -		vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
>> -	}
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
>> -		vcpu->arch.pfault_token = kvm_run->s.regs.pft;
>> -		vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
>> -		vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
>> +	riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
>> +	gscb = (struct gs_cb *) &run->s.regs.gscb;
>> +	vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
>> +	vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
>> +	if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> +		vcpu->arch.sie_block->todpr = run->s.regs.todpr;
>> +		vcpu->arch.sie_block->pp = run->s.regs.pp;
>> +		vcpu->arch.sie_block->gbea = run->s.regs.gbea;
>> +	}
>> +	if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
>> +		vcpu->arch.pfault_token = run->s.regs.pft;
>> +		vcpu->arch.pfault_select = run->s.regs.pfs;
>> +		vcpu->arch.pfault_compare = run->s.regs.pfc;
>>  		if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
>>  			kvm_clear_async_pf_completion_queue(vcpu);
>>  	}
>> @@ -4198,7 +4199,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>  	 * If userspace sets the riccb (e.g. after migration) to a valid state,
>>  	 * we should enable RI here instead of doing the lazy enablement.
>>  	 */
>> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>> +	if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>>  	    test_kvm_facility(vcpu->kvm, 64) &&
>>  	    riccb->v &&
>>  	    !(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
>> @@ -4209,7 +4210,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>  	 * If userspace sets the gscb (e.g. after migration) to non-zero,
>>  	 * we should enable GS here instead of doing the lazy enablement.
>>  	 */
>> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>> +	if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>>  	    test_kvm_facility(vcpu->kvm, 133) &&
>>  	    gscb->gssm &&
>>  	    !vcpu->arch.gs_enabled) {
>> @@ -4218,10 +4219,10 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>  		vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
>>  		vcpu->arch.gs_enabled = 1;
>>  	}
>> -	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>> +	if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>>  	    test_kvm_facility(vcpu->kvm, 82)) {
>>  		vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
>> -		vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
>> +		vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
>>  	}
>>  	if (MACHINE_HAS_GS) {
>>  		preempt_disable();
>> @@ -4232,45 +4233,47 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>  		}
>>  		if (vcpu->arch.gs_enabled) {
>>  			current->thread.gs_cb = (struct gs_cb *)
>> -						&vcpu->run->s.regs.gscb;
>> +						&run->s.regs.gscb;
>>  			restore_gs_cb(current->thread.gs_cb);
>>  		}
>>  		preempt_enable();
>>  	}
>> -	/* SIE will load etoken directly from SDNX and therefore kvm_run */
>> +	/* SIE will load etoken directly from SDNX and therefore run */
>>  }
>>  
>> -static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void sync_regs(struct kvm_vcpu *vcpu)
>>  {
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
>> -		kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
>> -		memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
>> +	struct kvm_run *run = vcpu->run;
>> +
>> +	if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
>> +		kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
>> +	if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
>> +		memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
>>  		/* some control register changes require a tlb flush */
>>  		kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
>>  	}
>> -	if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> -		kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
>> -		vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
>> +	if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> +		kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
>> +		vcpu->arch.sie_block->ckc = run->s.regs.ckc;
>>  	}
>>  	save_access_regs(vcpu->arch.host_acrs);
>> -	restore_access_regs(vcpu->run->s.regs.acrs);
>> +	restore_access_regs(run->s.regs.acrs);
>>  	/* save host (userspace) fprs/vrs */
>>  	save_fpu_regs();
>>  	vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
>>  	vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
>>  	if (MACHINE_HAS_VX)
>> -		current->thread.fpu.regs = vcpu->run->s.regs.vrs;
>> +		current->thread.fpu.regs = run->s.regs.vrs;
>>  	else
>> -		current->thread.fpu.regs = vcpu->run->s.regs.fprs;
>> -	current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
>> +		current->thread.fpu.regs = run->s.regs.fprs;
>> +	current->thread.fpu.fpc = run->s.regs.fpc;
>>  	if (test_fp_ctl(current->thread.fpu.fpc))
>>  		/* User space provided an invalid FPC, let's clear it */
>>  		current->thread.fpu.fpc = 0;
>>  
>>  	/* Sync fmt2 only data */
>>  	if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
>> -		sync_regs_fmt2(vcpu, kvm_run);
>> +		sync_regs_fmt2(vcpu);
>>  	} else {
>>  		/*
>>  		 * In several places we have to modify our internal view to
>> @@ -4282,19 +4285,21 @@ static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>  		 * do only accept the condition code from userspace.
>>  		 */
>>  		vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
>> -		vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
>> +		vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
>>  						   PSW_MASK_CC;
>>  	}
>>  
>> -	kvm_run->kvm_dirty_regs = 0;
>> +	run->kvm_dirty_regs = 0;
>>  }
>>  
>> -static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void store_regs_fmt2(struct kvm_vcpu *vcpu)
>>  {
>> -	kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
>> -	kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
>> -	kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
>> -	kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>> +	struct kvm_run *run = vcpu->run;
>> +
>> +	run->s.regs.todpr = vcpu->arch.sie_block->todpr;
>> +	run->s.regs.pp = vcpu->arch.sie_block->pp;
>> +	run->s.regs.gbea = vcpu->arch.sie_block->gbea;
>> +	run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>>  	if (MACHINE_HAS_GS) {
>>  		__ctl_set_bit(2, 4);
>>  		if (vcpu->arch.gs_enabled)
>> @@ -4310,39 +4315,41 @@ static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>>  	/* SIE will save etoken directly into SDNX and therefore kvm_run */
>>  }
>>  
>> -static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void store_regs(struct kvm_vcpu *vcpu)
>>  {
>> -	kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
>> -	kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
>> -	kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
>> -	memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
>> -	kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
>> -	kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
>> -	kvm_run->s.regs.pft = vcpu->arch.pfault_token;
>> -	kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
>> -	kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
>> -	save_access_regs(vcpu->run->s.regs.acrs);
>> +	struct kvm_run *run = vcpu->run;
>> +
>> +	run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
>> +	run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
>> +	run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
>> +	memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
>> +	run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
>> +	run->s.regs.ckc = vcpu->arch.sie_block->ckc;
>> +	run->s.regs.pft = vcpu->arch.pfault_token;
>> +	run->s.regs.pfs = vcpu->arch.pfault_select;
>> +	run->s.regs.pfc = vcpu->arch.pfault_compare;
>> +	save_access_regs(run->s.regs.acrs);
>>  	restore_access_regs(vcpu->arch.host_acrs);
>>  	/* Save guest register state */
>>  	save_fpu_regs();
>> -	vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
>> +	run->s.regs.fpc = current->thread.fpu.fpc;
>>  	/* Restore will be done lazily at return */
>>  	current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
>>  	current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
>>  	if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
>> -		store_regs_fmt2(vcpu, kvm_run);
>> +		store_regs_fmt2(vcpu);
>>  }
>>  
>>  int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>>  {
>> -	struct kvm_run *kvm_run = vcpu->run;
>> +	struct kvm_run *run = vcpu->run;
>>  	int rc;
>>  
>> -	if (kvm_run->immediate_exit)
>> +	if (run->immediate_exit)
>>  		return -EINTR;
>>  
>> -	if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
>> -	    kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>> +	if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
>> +	    run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>>  		return -EINVAL;
>>  
>>  	vcpu_load(vcpu);
>> @@ -4368,14 +4375,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>>  		goto out;
>>  	}
>>  
>> -	sync_regs(vcpu, kvm_run);
>> +	sync_regs(vcpu);
>>  	enable_cpu_timer_accounting(vcpu);
>>  
>>  	might_fault();
>>  	rc = __vcpu_run(vcpu);
>>  
>>  	if (signal_pending(current) && !rc) {
>> -		kvm_run->exit_reason = KVM_EXIT_INTR;
>> +		run->exit_reason = KVM_EXIT_INTR;
>>  		rc = -EINTR;
>>  	}
>>  
>> @@ -4390,7 +4397,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>>  	}
>>  
>>  	disable_cpu_timer_accounting(vcpu);
>> -	store_regs(vcpu, kvm_run);
>> +	store_regs(vcpu);
>>  
>>  	kvm_sigset_deactivate(vcpu);
>>  
>>
diff mbox series

Patch

diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
index 443af3ead739..cf420d013ba3 100644
--- a/arch/s390/kvm/kvm-s390.c
+++ b/arch/s390/kvm/kvm-s390.c
@@ -4173,24 +4173,25 @@  static int __vcpu_run(struct kvm_vcpu *vcpu)
 	return rc;
 }
 
-static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
+static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
 {
+	struct kvm_run *run = vcpu->run;
 	struct runtime_instr_cb *riccb;
 	struct gs_cb *gscb;
 
-	riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
-	gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
-	vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
-	vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
-	if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
-		vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
-		vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
-		vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
-	}
-	if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
-		vcpu->arch.pfault_token = kvm_run->s.regs.pft;
-		vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
-		vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
+	riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
+	gscb = (struct gs_cb *) &run->s.regs.gscb;
+	vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
+	vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
+	if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
+		vcpu->arch.sie_block->todpr = run->s.regs.todpr;
+		vcpu->arch.sie_block->pp = run->s.regs.pp;
+		vcpu->arch.sie_block->gbea = run->s.regs.gbea;
+	}
+	if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
+		vcpu->arch.pfault_token = run->s.regs.pft;
+		vcpu->arch.pfault_select = run->s.regs.pfs;
+		vcpu->arch.pfault_compare = run->s.regs.pfc;
 		if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
 			kvm_clear_async_pf_completion_queue(vcpu);
 	}
@@ -4198,7 +4199,7 @@  static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 	 * If userspace sets the riccb (e.g. after migration) to a valid state,
 	 * we should enable RI here instead of doing the lazy enablement.
 	 */
-	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
+	if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
 	    test_kvm_facility(vcpu->kvm, 64) &&
 	    riccb->v &&
 	    !(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
@@ -4209,7 +4210,7 @@  static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 	 * If userspace sets the gscb (e.g. after migration) to non-zero,
 	 * we should enable GS here instead of doing the lazy enablement.
 	 */
-	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
+	if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
 	    test_kvm_facility(vcpu->kvm, 133) &&
 	    gscb->gssm &&
 	    !vcpu->arch.gs_enabled) {
@@ -4218,10 +4219,10 @@  static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 		vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
 		vcpu->arch.gs_enabled = 1;
 	}
-	if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
+	if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
 	    test_kvm_facility(vcpu->kvm, 82)) {
 		vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
-		vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
+		vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
 	}
 	if (MACHINE_HAS_GS) {
 		preempt_disable();
@@ -4232,45 +4233,47 @@  static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 		}
 		if (vcpu->arch.gs_enabled) {
 			current->thread.gs_cb = (struct gs_cb *)
-						&vcpu->run->s.regs.gscb;
+						&run->s.regs.gscb;
 			restore_gs_cb(current->thread.gs_cb);
 		}
 		preempt_enable();
 	}
-	/* SIE will load etoken directly from SDNX and therefore kvm_run */
+	/* SIE will load etoken directly from SDNX and therefore run */
 }
 
-static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
+static void sync_regs(struct kvm_vcpu *vcpu)
 {
-	if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
-		kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
-	if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
-		memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
+	struct kvm_run *run = vcpu->run;
+
+	if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
+		kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
+	if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
+		memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
 		/* some control register changes require a tlb flush */
 		kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
 	}
-	if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
-		kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
-		vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
+	if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
+		kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
+		vcpu->arch.sie_block->ckc = run->s.regs.ckc;
 	}
 	save_access_regs(vcpu->arch.host_acrs);
-	restore_access_regs(vcpu->run->s.regs.acrs);
+	restore_access_regs(run->s.regs.acrs);
 	/* save host (userspace) fprs/vrs */
 	save_fpu_regs();
 	vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
 	vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
 	if (MACHINE_HAS_VX)
-		current->thread.fpu.regs = vcpu->run->s.regs.vrs;
+		current->thread.fpu.regs = run->s.regs.vrs;
 	else
-		current->thread.fpu.regs = vcpu->run->s.regs.fprs;
-	current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
+		current->thread.fpu.regs = run->s.regs.fprs;
+	current->thread.fpu.fpc = run->s.regs.fpc;
 	if (test_fp_ctl(current->thread.fpu.fpc))
 		/* User space provided an invalid FPC, let's clear it */
 		current->thread.fpu.fpc = 0;
 
 	/* Sync fmt2 only data */
 	if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
-		sync_regs_fmt2(vcpu, kvm_run);
+		sync_regs_fmt2(vcpu);
 	} else {
 		/*
 		 * In several places we have to modify our internal view to
@@ -4282,19 +4285,21 @@  static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 		 * do only accept the condition code from userspace.
 		 */
 		vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
-		vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
+		vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
 						   PSW_MASK_CC;
 	}
 
-	kvm_run->kvm_dirty_regs = 0;
+	run->kvm_dirty_regs = 0;
 }
 
-static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
+static void store_regs_fmt2(struct kvm_vcpu *vcpu)
 {
-	kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
-	kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
-	kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
-	kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
+	struct kvm_run *run = vcpu->run;
+
+	run->s.regs.todpr = vcpu->arch.sie_block->todpr;
+	run->s.regs.pp = vcpu->arch.sie_block->pp;
+	run->s.regs.gbea = vcpu->arch.sie_block->gbea;
+	run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
 	if (MACHINE_HAS_GS) {
 		__ctl_set_bit(2, 4);
 		if (vcpu->arch.gs_enabled)
@@ -4310,39 +4315,41 @@  static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 	/* SIE will save etoken directly into SDNX and therefore kvm_run */
 }
 
-static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
+static void store_regs(struct kvm_vcpu *vcpu)
 {
-	kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
-	kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
-	kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
-	memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
-	kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
-	kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
-	kvm_run->s.regs.pft = vcpu->arch.pfault_token;
-	kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
-	kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
-	save_access_regs(vcpu->run->s.regs.acrs);
+	struct kvm_run *run = vcpu->run;
+
+	run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
+	run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
+	run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
+	memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
+	run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
+	run->s.regs.ckc = vcpu->arch.sie_block->ckc;
+	run->s.regs.pft = vcpu->arch.pfault_token;
+	run->s.regs.pfs = vcpu->arch.pfault_select;
+	run->s.regs.pfc = vcpu->arch.pfault_compare;
+	save_access_regs(run->s.regs.acrs);
 	restore_access_regs(vcpu->arch.host_acrs);
 	/* Save guest register state */
 	save_fpu_regs();
-	vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
+	run->s.regs.fpc = current->thread.fpu.fpc;
 	/* Restore will be done lazily at return */
 	current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
 	current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
 	if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
-		store_regs_fmt2(vcpu, kvm_run);
+		store_regs_fmt2(vcpu);
 }
 
 int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
 {
-	struct kvm_run *kvm_run = vcpu->run;
+	struct kvm_run *run = vcpu->run;
 	int rc;
 
-	if (kvm_run->immediate_exit)
+	if (run->immediate_exit)
 		return -EINTR;
 
-	if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
-	    kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
+	if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
+	    run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
 		return -EINVAL;
 
 	vcpu_load(vcpu);
@@ -4368,14 +4375,14 @@  int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
 		goto out;
 	}
 
-	sync_regs(vcpu, kvm_run);
+	sync_regs(vcpu);
 	enable_cpu_timer_accounting(vcpu);
 
 	might_fault();
 	rc = __vcpu_run(vcpu);
 
 	if (signal_pending(current) && !rc) {
-		kvm_run->exit_reason = KVM_EXIT_INTR;
+		run->exit_reason = KVM_EXIT_INTR;
 		rc = -EINTR;
 	}
 
@@ -4390,7 +4397,7 @@  int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
 	}
 
 	disable_cpu_timer_accounting(vcpu);
-	store_regs(vcpu, kvm_run);
+	store_regs(vcpu);
 
 	kvm_sigset_deactivate(vcpu);