On 10/02/16 12:45, Christoffer Dall wrote:
> On Mon, Feb 08, 2016 at 11:40:21AM +0000, Marc Zyngier wrote:
>> GICv2 registers are *slow*. As in "terrifyingly slow". Which is bad.
>> But we're equaly bad, as we make a point in accessing them even if
>> we don't have any interrupt in flight.
>>
>> A good solution is to first find out if we have anything useful to
>> write into the GIC, and if we don't, to simply not do it. This
>> involves tracking which LRs actually have something valid there.
>>
>> Signed-off-by: Marc Zyngier <marc.zyng...@arm.com>
>> ---
>>  arch/arm64/kvm/hyp/vgic-v2-sr.c | 71 
>> ++++++++++++++++++++++++++++-------------
>>  include/kvm/arm_vgic.h          |  2 ++
>>  2 files changed, 51 insertions(+), 22 deletions(-)
>>
>> diff --git a/arch/arm64/kvm/hyp/vgic-v2-sr.c 
>> b/arch/arm64/kvm/hyp/vgic-v2-sr.c
>> index e717612..874a08d 100644
>> --- a/arch/arm64/kvm/hyp/vgic-v2-sr.c
>> +++ b/arch/arm64/kvm/hyp/vgic-v2-sr.c
>> @@ -38,28 +38,40 @@ void __hyp_text __vgic_v2_save_state(struct kvm_vcpu 
>> *vcpu)
>>  
>>      nr_lr = vcpu->arch.vgic_cpu.nr_lr;
>>      cpu_if->vgic_vmcr = readl_relaxed(base + GICH_VMCR);
>> -    cpu_if->vgic_misr = readl_relaxed(base + GICH_MISR);
>> -    eisr0  = readl_relaxed(base + GICH_EISR0);
>> -    elrsr0 = readl_relaxed(base + GICH_ELRSR0);
>> -    if (unlikely(nr_lr > 32)) {
>> -            eisr1  = readl_relaxed(base + GICH_EISR1);
>> -            elrsr1 = readl_relaxed(base + GICH_ELRSR1);
>> -    } else {
>> -            eisr1 = elrsr1 = 0;
>> -    }
>> +
>> +    if (vcpu->arch.vgic_cpu.live_lrs) {
>> +            eisr0  = readl_relaxed(base + GICH_EISR0);
>> +            elrsr0 = readl_relaxed(base + GICH_ELRSR0);
>> +            cpu_if->vgic_misr = readl_relaxed(base + GICH_MISR);
>> +            cpu_if->vgic_apr    = readl_relaxed(base + GICH_APR);
>> +
>> +            if (unlikely(nr_lr > 32)) {
>> +                    eisr1  = readl_relaxed(base + GICH_EISR1);
>> +                    elrsr1 = readl_relaxed(base + GICH_ELRSR1);
>> +            } else {
>> +                    eisr1 = elrsr1 = 0;
>> +            }
>> +
>>  #ifdef CONFIG_CPU_BIG_ENDIAN
>> -    cpu_if->vgic_eisr  = ((u64)eisr0 << 32) | eisr1;
>> -    cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1;
>> +            cpu_if->vgic_eisr  = ((u64)eisr0 << 32) | eisr1;
>> +            cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1;
>>  #else
>> -    cpu_if->vgic_eisr  = ((u64)eisr1 << 32) | eisr0;
>> -    cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0;
>> +            cpu_if->vgic_eisr  = ((u64)eisr1 << 32) | eisr0;
>> +            cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0;
>>  #endif
>> -    cpu_if->vgic_apr    = readl_relaxed(base + GICH_APR);
>>  
>> -    writel_relaxed(0, base + GICH_HCR);
>> +            for (i = 0; i < nr_lr; i++)
>> +                    if (vcpu->arch.vgic_cpu.live_lrs & (1UL << i))
>> +                            cpu_if->vgic_lr[i] = readl_relaxed(base + 
>> GICH_LR0 + (i * 4));
>>  
>> -    for (i = 0; i < nr_lr; i++)
>> -            cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4));
>> +            writel_relaxed(0, base + GICH_HCR);
>> +
>> +            vcpu->arch.vgic_cpu.live_lrs = 0;
>> +    } else {
>> +            cpu_if->vgic_eisr = 0;
>> +            cpu_if->vgic_elrsr = ~0UL;
>> +            cpu_if->vgic_misr = 0;
>> +    }
>>  }
>>  
>>  /* vcpu is already in the HYP VA space */
>> @@ -70,15 +82,30 @@ void __hyp_text __vgic_v2_restore_state(struct kvm_vcpu 
>> *vcpu)
>>      struct vgic_dist *vgic = &kvm->arch.vgic;
>>      void __iomem *base = kern_hyp_va(vgic->vctrl_base);
>>      int i, nr_lr;
>> +    u64 live_lrs = 0;
>>  
>>      if (!base)
>>              return;
>>  
>> -    writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR);
>> -    writel_relaxed(cpu_if->vgic_vmcr, base + GICH_VMCR);
>> -    writel_relaxed(cpu_if->vgic_apr, base + GICH_APR);
>> -
>>      nr_lr = vcpu->arch.vgic_cpu.nr_lr;
>> +
>>      for (i = 0; i < nr_lr; i++)
>> -            writel_relaxed(cpu_if->vgic_lr[i], base + GICH_LR0 + (i * 4));
>> +            if (cpu_if->vgic_lr[i] & GICH_LR_STATE)
>> +                    live_lrs |= 1UL << i;
>> +
>> +    if (live_lrs) {
>> +            writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR);
>> +            writel_relaxed(cpu_if->vgic_apr, base + GICH_APR);
>> +            for (i = 0; i < nr_lr; i++) {
>> +                    u32 val = 0;
>> +
>> +                    if (live_lrs & (1UL << i))
>> +                            val = cpu_if->vgic_lr[i];
>> +
>> +                    writel_relaxed(val, base + GICH_LR0 + (i * 4));
>> +            }
>> +    }
>> +
>> +    writel_relaxed(cpu_if->vgic_vmcr, base + GICH_VMCR);
> 
> couldn't you optimize this out by storing the last read value and
> compare if anything changed?  (you'd have to invalidate the cached value
> on vcpu_put obviously).

Yeah, very good point. Only the guest can update this, so we could even
move it to vcpu_load/vcpu_put entirely, and never save/restore it inside
the run loop.

I'll keep that for a further patch, as it requires a bit of infrastructure.

Thanks,

        M.
-- 
Jazz is not dead. It just smells funny...
_______________________________________________
kvmarm mailing list
kvmarm@lists.cs.columbia.edu
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm

Reply via email to