On Fri, Jan 25, 2019 at 05:53:47PM +0800, Ming Lei wrote:
> Now allocating interrupt sets can be done via .setup_affinity()
> easily, so remove the support for allocating interrupt sets.
> 
> With this change, we don't need the limit of 'minvec == maxvec'
> any more in pci_alloc_irq_vectors_affinity().
> 
> Meantime irq_create_affinity_masks() gets simplified a lot.
> 
> Signed-off-by: Ming Lei <ming....@redhat.com>

Acked-by: Bjorn Helgaas <bhelg...@google.com>   # pci/msi.c parts

> ---
>  drivers/pci/msi.c         | 14 -------------
>  include/linux/interrupt.h |  4 ----
>  kernel/irq/affinity.c     | 52 
> +++++++++++------------------------------------
>  3 files changed, 12 insertions(+), 58 deletions(-)
> 
> diff --git a/drivers/pci/msi.c b/drivers/pci/msi.c
> index 4c0b47867258..331483de1294 100644
> --- a/drivers/pci/msi.c
> +++ b/drivers/pci/msi.c
> @@ -1035,13 +1035,6 @@ static int __pci_enable_msi_range(struct pci_dev *dev, 
> int minvec, int maxvec,
>       if (maxvec < minvec)
>               return -ERANGE;
>  
> -     /*
> -      * If the caller is passing in sets, we can't support a range of
> -      * vectors. The caller needs to handle that.
> -      */
> -     if (affd && affd->nr_sets && minvec != maxvec)
> -             return -EINVAL;
> -
>       if (WARN_ON_ONCE(dev->msi_enabled))
>               return -EINVAL;
>  
> @@ -1093,13 +1086,6 @@ static int __pci_enable_msix_range(struct pci_dev *dev,
>       if (maxvec < minvec)
>               return -ERANGE;
>  
> -     /*
> -      * If the caller is passing in sets, we can't support a range of
> -      * supported vectors. The caller needs to handle that.
> -      */
> -     if (affd && affd->nr_sets && minvec != maxvec)
> -             return -EINVAL;
> -
>       if (WARN_ON_ONCE(dev->msix_enabled))
>               return -EINVAL;
>  
> diff --git a/include/linux/interrupt.h b/include/linux/interrupt.h
> index b820b07f3b55..a035e165f405 100644
> --- a/include/linux/interrupt.h
> +++ b/include/linux/interrupt.h
> @@ -260,8 +260,6 @@ struct irq_affinity_desc {
>   *                   and driver has to handle pre_vectors & post_vectors
>   *                   correctly, set 'is_managed' flag correct too
>   * @priv:            Private data of @setup_affinity
> - * @nr_sets:         Length of passed in *sets array
> - * @sets:            Number of affinitized sets
>   */
>  struct irq_affinity {
>       int     pre_vectors;
> @@ -270,8 +268,6 @@ struct irq_affinity {
>                                 struct irq_affinity_desc *,
>                                 unsigned int);
>       void    *priv;
> -     int     nr_sets;
> -     int     *sets;
>  };
>  
>  #if defined(CONFIG_SMP)
> diff --git a/kernel/irq/affinity.c b/kernel/irq/affinity.c
> index 524fdcda9f85..e8fea65325d9 100644
> --- a/kernel/irq/affinity.c
> +++ b/kernel/irq/affinity.c
> @@ -269,9 +269,9 @@ struct irq_affinity_desc *
>  irq_create_affinity_masks(int nvecs, const struct irq_affinity *affd)
>  {
>       int affvecs = nvecs - affd->pre_vectors - affd->post_vectors;
> -     int curvec, usedvecs;
> +     int curvec;
>       struct irq_affinity_desc *masks = NULL;
> -     int i, nr_sets;
> +     int i;
>  
>       /*
>        * If there aren't any vectors left after applying the pre/post
> @@ -293,34 +293,14 @@ irq_create_affinity_masks(int nvecs, const struct 
> irq_affinity *affd)
>       /* Fill out vectors at the beginning that don't need affinity */
>       for (curvec = 0; curvec < affd->pre_vectors; curvec++)
>               cpumask_copy(&masks[curvec].mask, irq_default_affinity);
> -     /*
> -      * Spread on present CPUs starting from affd->pre_vectors. If we
> -      * have multiple sets, build each sets affinity mask separately.
> -      */
> -     nr_sets = affd->nr_sets;
> -     if (!nr_sets)
> -             nr_sets = 1;
> -
> -     for (i = 0, usedvecs = 0; i < nr_sets; i++) {
> -             int this_vecs = affd->sets ? affd->sets[i] : affvecs;
> -             int ret;
> -
> -             ret = irq_build_affinity_masks(affd, curvec, this_vecs,
> -                                             curvec, masks);
> -             if (ret) {
> -                     kfree(masks);
> -                     return NULL;
> -             }
> -             curvec += this_vecs;
> -             usedvecs += this_vecs;
> +
> +     if (irq_build_affinity_masks(affd, curvec, affvecs, curvec, masks)) {
> +             kfree(masks);
> +             return NULL;
>       }
>  
>       /* Fill out vectors at the end that don't need affinity */
> -     if (usedvecs >= affvecs)
> -             curvec = affd->pre_vectors + affvecs;
> -     else
> -             curvec = affd->pre_vectors + usedvecs;
> -     for (; curvec < nvecs; curvec++)
> +     for (curvec = affd->pre_vectors + affvecs; curvec < nvecs; curvec++)
>               cpumask_copy(&masks[curvec].mask, irq_default_affinity);
>  
>       /* Mark the managed interrupts */
> @@ -340,21 +320,13 @@ int irq_calc_affinity_vectors(int minvec, int maxvec, 
> const struct irq_affinity
>  {
>       int resv = affd->pre_vectors + affd->post_vectors;
>       int vecs = maxvec - resv;
> -     int set_vecs;
> +     int ret;
>  
>       if (resv > minvec)
>               return 0;
>  
> -     if (affd->nr_sets) {
> -             int i;
> -
> -             for (i = 0, set_vecs = 0;  i < affd->nr_sets; i++)
> -                     set_vecs += affd->sets[i];
> -     } else {
> -             get_online_cpus();
> -             set_vecs = cpumask_weight(cpu_possible_mask);
> -             put_online_cpus();
> -     }
> -
> -     return resv + min(set_vecs, vecs);
> +     get_online_cpus();
> +     ret = min_t(int, cpumask_weight(cpu_possible_mask), vecs) + resv;
> +     put_online_cpus();
> +     return ret;
>  }
> -- 
> 2.9.5
> 

Reply via email to