> From: Lu Baolu <baolu...@linux.intel.com>
> Sent: Monday, July 6, 2020 8:26 AM
> 
> A pasid might be bound to a page table from a VM guest via the iommu
> ops.sva_bind_gpasid. In this case, when a DMA page fault is detected
> on the physical IOMMU, we need to inject the page fault request into
> the guest. After the guest completes handling the page fault, a page
> response need to be sent back via the iommu ops.page_response().
> 
> This adds support to report a page request fault. Any external module
> which is interested in handling this fault should regiester a notifier
> callback.

be specific on which notifier to be registered...

> 
> Co-developed-by: Jacob Pan <jacob.jun....@linux.intel.com>
> Signed-off-by: Jacob Pan <jacob.jun....@linux.intel.com>
> Co-developed-by: Liu Yi L <yi.l....@intel.com>
> Signed-off-by: Liu Yi L <yi.l....@intel.com>
> Signed-off-by: Lu Baolu <baolu...@linux.intel.com>
> ---
>  drivers/iommu/intel/svm.c | 99 ++++++++++++++++++++++++++++++++-------
>  1 file changed, 81 insertions(+), 18 deletions(-)
> 
> diff --git a/drivers/iommu/intel/svm.c b/drivers/iommu/intel/svm.c
> index c23167877b2b..08c58c2b1a06 100644
> --- a/drivers/iommu/intel/svm.c
> +++ b/drivers/iommu/intel/svm.c
> @@ -815,6 +815,57 @@ static void intel_svm_drain_prq(struct device *dev,
> int pasid)
>       }
>  }
> 
> +static int prq_to_iommu_prot(struct page_req_dsc *req)
> +{
> +     int prot = 0;
> +
> +     if (req->rd_req)
> +             prot |= IOMMU_FAULT_PERM_READ;
> +     if (req->wr_req)
> +             prot |= IOMMU_FAULT_PERM_WRITE;
> +     if (req->exe_req)
> +             prot |= IOMMU_FAULT_PERM_EXEC;
> +     if (req->pm_req)
> +             prot |= IOMMU_FAULT_PERM_PRIV;
> +
> +     return prot;
> +}
> +
> +static int
> +intel_svm_prq_report(struct device *dev, struct page_req_dsc *desc)
> +{
> +     struct iommu_fault_event event;
> +     u8 bus, devfn;
> +
> +     memset(&event, 0, sizeof(struct iommu_fault_event));
> +     bus = PCI_BUS_NUM(desc->rid);
> +     devfn = desc->rid & 0xff;

not required.

> +
> +     /* Fill in event data for device specific processing */
> +     event.fault.type = IOMMU_FAULT_PAGE_REQ;
> +     event.fault.prm.addr = desc->addr;
> +     event.fault.prm.pasid = desc->pasid;
> +     event.fault.prm.grpid = desc->prg_index;
> +     event.fault.prm.perm = prq_to_iommu_prot(desc);
> +
> +     /*
> +      * Set last page in group bit if private data is present,
> +      * page response is required as it does for LPIG.
> +      */

move to priv_data_present check?

> +     if (desc->lpig)
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE;
> +     if (desc->pasid_present)
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_REQUEST_PASID_VALID;
> +     if (desc->priv_data_present) {
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE;
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_REQUEST_PRIV_DATA;
> +             memcpy(event.fault.prm.private_data, desc->priv_data,
> +                    sizeof(desc->priv_data));
> +     }
> +
> +     return iommu_report_device_fault(dev, &event);
> +}
> +
>  static irqreturn_t prq_event_thread(int irq, void *d)
>  {
>       struct intel_iommu *iommu = d;
> @@ -828,7 +879,7 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>       tail = dmar_readq(iommu->reg + DMAR_PQT_REG) &
> PRQ_RING_MASK;
>       head = dmar_readq(iommu->reg + DMAR_PQH_REG) &
> PRQ_RING_MASK;
>       while (head != tail) {
> -             struct intel_svm_dev *sdev;
> +             struct intel_svm_dev *sdev = NULL;

move to outside of the loop, otherwise later check always hit "if (!sdev)"

>               struct vm_area_struct *vma;
>               struct page_req_dsc *req;
>               struct qi_desc resp;
> @@ -864,6 +915,20 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>                       }
>               }
> 
> +             if (!sdev || sdev->sid != req->rid) {
> +                     struct intel_svm_dev *t;
> +
> +                     sdev = NULL;
> +                     rcu_read_lock();
> +                     list_for_each_entry_rcu(t, &svm->devs, list) {
> +                             if (t->sid == req->rid) {
> +                                     sdev = t;
> +                                     break;
> +                             }
> +                     }
> +                     rcu_read_unlock();
> +             }
> +
>               result = QI_RESP_INVALID;
>               /* Since we're using init_mm.pgd directly, we should never
> take
>                * any faults on kernel addresses. */
> @@ -874,6 +939,17 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>               if (!is_canonical_address(address))
>                       goto bad_req;
> 
> +             /*
> +              * If prq is to be handled outside iommu driver via receiver of
> +              * the fault notifiers, we skip the page response here.
> +              */
> +             if (svm->flags & SVM_FLAG_GUEST_MODE) {
> +                     if (sdev && !intel_svm_prq_report(sdev->dev, req))
> +                             goto prq_advance;
> +                     else
> +                             goto bad_req;
> +             }
> +
>               /* If the mm is already defunct, don't handle faults. */
>               if (!mmget_not_zero(svm->mm))
>                       goto bad_req;
> @@ -892,24 +968,10 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>                       goto invalid;
> 
>               result = QI_RESP_SUCCESS;
> -     invalid:
> +invalid:
>               mmap_read_unlock(svm->mm);
>               mmput(svm->mm);
> -     bad_req:
> -             /* Accounting for major/minor faults? */
> -             rcu_read_lock();
> -             list_for_each_entry_rcu(sdev, &svm->devs, list) {
> -                     if (sdev->sid == req->rid)
> -                             break;
> -             }
> -             /* Other devices can go away, but the drivers are not
> permitted
> -              * to unbind while any page faults might be in flight. So it's
> -              * OK to drop the 'lock' here now we have it. */

should we keep and move this comment to earlier sdev lookup? and
regarding to guest unbind, ae we preventing the fault owner (outside
of iommu driver) to unbind against in-flight fault request?

> -             rcu_read_unlock();
> -
> -             if (WARN_ON(&sdev->list == &svm->devs))
> -                     sdev = NULL;

similarly should we keep the WARN_ON check here?

> -
> +bad_req:
>               if (sdev && sdev->ops && sdev->ops->fault_cb) {
>                       int rwxp = (req->rd_req << 3) | (req->wr_req << 2) |
>                               (req->exe_req << 1) | (req->pm_req);
> @@ -920,7 +982,7 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>                  and these can be NULL. Do not use them below this point!
> */
>               sdev = NULL;
>               svm = NULL;
> -     no_pasid:
> +no_pasid:
>               if (req->lpig || req->priv_data_present) {
>                       /*
>                        * Per VT-d spec. v3.0 ch7.7, system software must
> @@ -945,6 +1007,7 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>                       resp.qw3 = 0;
>                       qi_submit_sync(iommu, &resp, 1, 0);
>               }
> +prq_advance:
>               head = (head + sizeof(*req)) & PRQ_RING_MASK;
>       }
> 
> --
> 2.17.1

Thanks
Kevin

Reply via email to