Level sensitive interrupts are exposed as maskable and automasked
interrupts and are masked and disabled automatically when they fire.

Signed-off-by: Antonios Motakis <a.mota...@virtualopensystems.com>
---
 drivers/vfio/platform/vfio_platform_irq.c     | 99 ++++++++++++++++++++++++++-
 drivers/vfio/platform/vfio_platform_private.h |  2 +
 2 files changed, 98 insertions(+), 3 deletions(-)

diff --git a/drivers/vfio/platform/vfio_platform_irq.c 
b/drivers/vfio/platform/vfio_platform_irq.c
index 4896581..b3602cb 100644
--- a/drivers/vfio/platform/vfio_platform_irq.c
+++ b/drivers/vfio/platform/vfio_platform_irq.c
@@ -23,12 +23,59 @@
 
 #include "vfio_platform_private.h"
 
+static void vfio_platform_mask(struct vfio_platform_irq *irq_ctx)
+{
+       unsigned long flags;
+
+       spin_lock_irqsave(&irq_ctx->lock, flags);
+
+       if (!irq_ctx->masked) {
+               disable_irq_nosync(irq_ctx->hwirq);
+               irq_ctx->masked = true;
+       }
+
+       spin_unlock_irqrestore(&irq_ctx->lock, flags);
+}
+
 static int vfio_platform_set_irq_mask(struct vfio_platform_device *vdev,
                                      unsigned index, unsigned start,
                                      unsigned count, uint32_t flags,
                                      void *data)
 {
-       return -EINVAL;
+       if (start != 0 || count != 1)
+               return -EINVAL;
+
+       if (!(vdev->irqs[index].flags & VFIO_IRQ_INFO_MASKABLE))
+               return -EINVAL;
+
+       if (flags & VFIO_IRQ_SET_DATA_EVENTFD)
+               return -EINVAL; /* not implemented yet */
+
+       if (flags & VFIO_IRQ_SET_DATA_NONE) {
+               vfio_platform_mask(&vdev->irqs[index]);
+
+       } else if (flags & VFIO_IRQ_SET_DATA_BOOL) {
+               uint8_t mask = *(uint8_t *)data;
+
+               if (mask)
+                       vfio_platform_mask(&vdev->irqs[index]);
+       }
+
+       return 0;
+}
+
+static void vfio_platform_unmask(struct vfio_platform_irq *irq_ctx)
+{
+       unsigned long flags;
+
+       spin_lock_irqsave(&irq_ctx->lock, flags);
+
+       if (irq_ctx->masked) {
+               enable_irq(irq_ctx->hwirq);
+               irq_ctx->masked = false;
+       }
+
+       spin_unlock_irqrestore(&irq_ctx->lock, flags);
 }
 
 static int vfio_platform_set_irq_unmask(struct vfio_platform_device *vdev,
@@ -36,7 +83,50 @@ static int vfio_platform_set_irq_unmask(struct 
vfio_platform_device *vdev,
                                        unsigned count, uint32_t flags,
                                        void *data)
 {
-       return -EINVAL;
+       if (start != 0 || count != 1)
+               return -EINVAL;
+
+       if (!(vdev->irqs[index].flags & VFIO_IRQ_INFO_MASKABLE))
+               return -EINVAL;
+
+       if (flags & VFIO_IRQ_SET_DATA_EVENTFD)
+               return -EINVAL; /* not implemented yet */
+
+       if (flags & VFIO_IRQ_SET_DATA_NONE) {
+               vfio_platform_unmask(&vdev->irqs[index]);
+
+       } else if (flags & VFIO_IRQ_SET_DATA_BOOL) {
+               uint8_t unmask = *(uint8_t *)data;
+
+               if (unmask)
+                       vfio_platform_unmask(&vdev->irqs[index]);
+       }
+
+       return 0;
+}
+
+static irqreturn_t vfio_automasked_irq_handler(int irq, void *dev_id)
+{
+       struct vfio_platform_irq *irq_ctx = dev_id;
+       unsigned long flags;
+       int ret = IRQ_NONE;
+
+       spin_lock_irqsave(&irq_ctx->lock, flags);
+
+       if (!irq_ctx->masked) {
+               ret = IRQ_HANDLED;
+
+               /* automask maskable interrupts */
+               disable_irq_nosync(irq_ctx->hwirq);
+               irq_ctx->masked = true;
+       }
+
+       spin_unlock_irqrestore(&irq_ctx->lock, flags);
+
+       if (ret == IRQ_HANDLED)
+               eventfd_signal(irq_ctx->trigger, 1);
+
+       return ret;
 }
 
 static irqreturn_t vfio_irq_handler(int irq, void *dev_id)
@@ -106,7 +196,7 @@ static int vfio_platform_set_irq_trigger(struct 
vfio_platform_device *vdev,
        irq_handler_t handler;
 
        if (vdev->irqs[index].flags & VFIO_IRQ_INFO_AUTOMASKED)
-               return -EINVAL; /* not implemented */
+               handler = vfio_automasked_irq_handler;
        else
                handler = vfio_irq_handler;
 
@@ -178,6 +268,8 @@ int vfio_platform_irq_init(struct vfio_platform_device 
*vdev)
                if (hwirq < 0)
                        goto err;
 
+               spin_lock_init(&vdev->irqs[i].lock);
+
                vdev->irqs[i].flags = VFIO_IRQ_INFO_EVENTFD;
 
                if (irq_get_trigger_type(hwirq) & IRQ_TYPE_LEVEL_MASK)
@@ -186,6 +278,7 @@ int vfio_platform_irq_init(struct vfio_platform_device 
*vdev)
 
                vdev->irqs[i].count = 1;
                vdev->irqs[i].hwirq = hwirq;
+               vdev->irqs[i].masked = false;
        }
 
        vdev->num_irqs = cnt;
diff --git a/drivers/vfio/platform/vfio_platform_private.h 
b/drivers/vfio/platform/vfio_platform_private.h
index b705f17..eef6d1b 100644
--- a/drivers/vfio/platform/vfio_platform_private.h
+++ b/drivers/vfio/platform/vfio_platform_private.h
@@ -30,6 +30,8 @@ struct vfio_platform_irq {
        int                     hwirq;
        char                    *name;
        struct eventfd_ctx      *trigger;
+       bool                    masked;
+       spinlock_t              lock;
 };
 
 struct vfio_platform_region {
-- 
2.1.4

_______________________________________________
iommu mailing list
iommu@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/iommu

Reply via email to