Ping, this patch is still good for most recent api-next branch

Best Regards, Yi

On 23 December 2016 at 10:32, Yi He <yi...@linaro.org> wrote:

> SP scheduler hangs in process mode performance test
> due to global data structure were not created in shared
> memory region.
>
> Signed-off-by: Yi He <yi...@linaro.org>
> ---
> since v1: rebased upon Petri's linux-gen: schedule_sp: use ring as
> priority queue
>
>  platform/linux-generic/odp_schedule_sp.c | 100
> ++++++++++++++++++-------------
>  1 file changed, 60 insertions(+), 40 deletions(-)
>
> diff --git a/platform/linux-generic/odp_schedule_sp.c
> b/platform/linux-generic/odp_schedule_sp.c
> index 5150d28..bb7416a 100644
> --- a/platform/linux-generic/odp_schedule_sp.c
> +++ b/platform/linux-generic/odp_schedule_sp.c
> @@ -9,6 +9,7 @@
>  #include <odp/api/thread.h>
>  #include <odp/api/time.h>
>  #include <odp/api/schedule.h>
> +#include <odp/api/shared_memory.h>
>  #include <odp_schedule_if.h>
>  #include <odp_debug_internal.h>
>  #include <odp_align_internal.h>
> @@ -108,6 +109,7 @@ typedef struct {
>         sched_cmd_t   pktio_cmd[NUM_PKTIO];
>         prio_queue_t  prio_queue[NUM_GROUP][NUM_PRIO];
>         sched_group_t sched_group;
> +       odp_shm_t     shm;
>  } sched_global_t;
>
>  typedef struct {
> @@ -119,7 +121,7 @@ typedef struct {
>         int          group[NUM_GROUP];
>  } sched_local_t;
>
> -static sched_global_t sched_global;
> +static sched_global_t *sched_global;
>  static __thread sched_local_t sched_local;
>
>  static inline uint32_t index_to_ring_idx(int pktio, uint32_t index)
> @@ -145,30 +147,44 @@ static inline uint32_t index_from_ring_idx(uint32_t
> *index, uint32_t ring_idx)
>  static int init_global(void)
>  {
>         int i, j;
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       odp_shm_t shm;
> +       sched_group_t *sched_group = NULL;
>
>         ODP_DBG("Using SP scheduler\n");
>
> -       memset(&sched_global, 0, sizeof(sched_global_t));
> +       shm = odp_shm_reserve("sp_scheduler",
> +                             sizeof(sched_global_t),
> +                             ODP_CACHE_LINE_SIZE, 0);
> +
> +       sched_global = odp_shm_addr(shm);
> +
> +       if (sched_global == NULL) {
> +               ODP_ERR("Schedule init: Shm reserve failed.\n");
> +               return -1;
> +       }
> +
> +       memset(sched_global, 0, sizeof(sched_global_t));
> +       sched_global->shm = shm;
>
>         for (i = 0; i < NUM_QUEUE; i++) {
> -               sched_global.queue_cmd[i].s.type     = CMD_QUEUE;
> -               sched_global.queue_cmd[i].s.index    = i;
> -               sched_global.queue_cmd[i].s.ring_idx =
> index_to_ring_idx(0, i);
> +               sched_global->queue_cmd[i].s.type     = CMD_QUEUE;
> +               sched_global->queue_cmd[i].s.index    = i;
> +               sched_global->queue_cmd[i].s.ring_idx =
> index_to_ring_idx(0, i);
>         }
>
>         for (i = 0; i < NUM_PKTIO; i++) {
> -               sched_global.pktio_cmd[i].s.type     = CMD_PKTIO;
> -               sched_global.pktio_cmd[i].s.index    = i;
> -               sched_global.pktio_cmd[i].s.ring_idx =
> index_to_ring_idx(1, i);
> -               sched_global.pktio_cmd[i].s.prio     = PKTIN_PRIO;
> -               sched_global.pktio_cmd[i].s.group    = GROUP_PKTIN;
> +               sched_global->pktio_cmd[i].s.type     = CMD_PKTIO;
> +               sched_global->pktio_cmd[i].s.index    = i;
> +               sched_global->pktio_cmd[i].s.ring_idx =
> index_to_ring_idx(1, i);
> +               sched_global->pktio_cmd[i].s.prio     = PKTIN_PRIO;
> +               sched_global->pktio_cmd[i].s.group    = GROUP_PKTIN;
>         }
>
>         for (i = 0; i < NUM_GROUP; i++)
>                 for (j = 0; j < NUM_PRIO; j++)
> -                       ring_init(&sched_global.prio_queue[i][j].ring);
> +                       ring_init(&sched_global->prio_queue[i][j].ring);
>
> +       sched_group = &sched_global->sched_group;
>         odp_ticketlock_init(&sched_group->s.lock);
>
>         for (i = 0; i < NUM_THREAD; i++)
> @@ -202,16 +218,22 @@ static int init_local(void)
>
>  static int term_global(void)
>  {
> -       int qi;
> +       int qi, ret = 0;
>
>         for (qi = 0; qi < NUM_QUEUE; qi++) {
> -               if (sched_global.queue_cmd[qi].s.init) {
> +               if (sched_global->queue_cmd[qi].s.init) {
>                         /* todo: dequeue until empty ? */
>                         sched_cb_queue_destroy_finalize(qi);
>                 }
>         }
>
> -       return 0;
> +       ret = odp_shm_free(sched_global->shm);
> +       if (ret < 0) {
> +               ODP_ERR("Shm free failed for sp_scheduler");
> +               ret = -1;
> +       }
> +
> +       return ret;
>  }
>
>  static int term_local(void)
> @@ -267,7 +289,7 @@ static void remove_group(sched_group_t *sched_group,
> int thr, int group)
>
>  static int thr_add(odp_schedule_group_t group, int thr)
>  {
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>
>         if (group < 0 || group >= NUM_GROUP)
>                 return -1;
> @@ -292,7 +314,7 @@ static int thr_add(odp_schedule_group_t group, int thr)
>
>  static int thr_rem(odp_schedule_group_t group, int thr)
>  {
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>
>         if (group < 0 || group >= NUM_GROUP)
>                 return -1;
> @@ -320,7 +342,7 @@ static int num_grps(void)
>
>  static int init_queue(uint32_t qi, const odp_schedule_param_t
> *sched_param)
>  {
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>         odp_schedule_group_t group = sched_param->group;
>         int prio = 0;
>
> @@ -333,18 +355,18 @@ static int init_queue(uint32_t qi, const
> odp_schedule_param_t *sched_param)
>         if (sched_param->prio > 0)
>                 prio = LOWEST_QUEUE_PRIO;
>
> -       sched_global.queue_cmd[qi].s.prio  = prio;
> -       sched_global.queue_cmd[qi].s.group = group;
> -       sched_global.queue_cmd[qi].s.init  = 1;
> +       sched_global->queue_cmd[qi].s.prio  = prio;
> +       sched_global->queue_cmd[qi].s.group = group;
> +       sched_global->queue_cmd[qi].s.init  = 1;
>
>         return 0;
>  }
>
>  static void destroy_queue(uint32_t qi)
>  {
> -       sched_global.queue_cmd[qi].s.prio  = 0;
> -       sched_global.queue_cmd[qi].s.group = 0;
> -       sched_global.queue_cmd[qi].s.init  = 0;
> +       sched_global->queue_cmd[qi].s.prio  = 0;
> +       sched_global->queue_cmd[qi].s.group = 0;
> +       sched_global->queue_cmd[qi].s.init  = 0;
>  }
>
>  static inline void add_tail(sched_cmd_t *cmd)
> @@ -354,8 +376,7 @@ static inline void add_tail(sched_cmd_t *cmd)
>         int prio     = cmd->s.prio;
>         uint32_t idx = cmd->s.ring_idx;
>
> -       prio_queue = &sched_global.prio_queue[group][prio];
> -
> +       prio_queue = &sched_global->prio_queue[group][prio];
>         ring_enq(&prio_queue->ring, RING_MASK, idx);
>  }
>
> @@ -365,8 +386,7 @@ static inline sched_cmd_t *rem_head(int group, int
> prio)
>         uint32_t ring_idx, index;
>         int pktio;
>
> -       prio_queue = &sched_global.prio_queue[group][prio];
> -
> +       prio_queue = &sched_global->prio_queue[group][prio];
>         ring_idx = ring_deq(&prio_queue->ring, RING_MASK);
>
>         if (ring_idx == RING_EMPTY)
> @@ -375,16 +395,16 @@ static inline sched_cmd_t *rem_head(int group, int
> prio)
>         pktio = index_from_ring_idx(&index, ring_idx);
>
>         if (pktio)
> -               return &sched_global.pktio_cmd[index];
> +               return &sched_global->pktio_cmd[index];
>
> -       return &sched_global.queue_cmd[index];
> +       return &sched_global->queue_cmd[index];
>  }
>
>  static int sched_queue(uint32_t qi)
>  {
>         sched_cmd_t *cmd;
>
> -       cmd = &sched_global.queue_cmd[qi];
> +       cmd = &sched_global->queue_cmd[qi];
>         add_tail(cmd);
>
>         return 0;
> @@ -410,7 +430,7 @@ static void pktio_start(int pktio_index, int num, int
> pktin_idx[])
>         ODP_DBG("pktio index: %i, %i pktin queues %i\n",
>                 pktio_index, num, pktin_idx[0]);
>
> -       cmd = &sched_global.pktio_cmd[pktio_index];
> +       cmd = &sched_global->pktio_cmd[pktio_index];
>
>         if (num > NUM_PKTIN)
>                 ODP_ABORT("Supports only %i pktin queues per interface\n",
> @@ -428,7 +448,7 @@ static inline sched_cmd_t *sched_cmd(void)
>  {
>         int prio, i;
>         int thr = sched_local.thr_id;
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>         thr_group_t *thr_group = &sched_group->s.thr[thr];
>         uint32_t gen_cnt;
>
> @@ -602,7 +622,7 @@ static odp_schedule_group_t
> schedule_group_create(const char *name,
>                                                   const odp_thrmask_t
> *thrmask)
>  {
>         odp_schedule_group_t group = ODP_SCHED_GROUP_INVALID;
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>         int i;
>
>         odp_ticketlock_lock(&sched_group->s.lock);
> @@ -633,7 +653,7 @@ static odp_schedule_group_t
> schedule_group_create(const char *name,
>
>  static int schedule_group_destroy(odp_schedule_group_t group)
>  {
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>
>         if (group < NUM_STATIC_GROUP || group >= NUM_GROUP)
>                 return -1;
> @@ -656,7 +676,7 @@ static int schedule_group_destroy(odp_schedule_group_t
> group)
>  static odp_schedule_group_t schedule_group_lookup(const char *name)
>  {
>         odp_schedule_group_t group = ODP_SCHED_GROUP_INVALID;
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>         int i;
>
>         odp_ticketlock_lock(&sched_group->s.lock);
> @@ -677,7 +697,7 @@ static int schedule_group_join(odp_schedule_group_t
> group,
>                                const odp_thrmask_t *thrmask)
>  {
>         int thr;
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>
>         if (group < 0 || group >= NUM_GROUP)
>                 return -1;
> @@ -709,7 +729,7 @@ static int schedule_group_leave(odp_schedule_group_t
> group,
>                                 const odp_thrmask_t *thrmask)
>  {
>         int thr;
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>         odp_thrmask_t *all = &sched_group->s.group[GROUP_ALL].mask;
>         odp_thrmask_t not;
>
> @@ -743,7 +763,7 @@ static int schedule_group_leave(odp_schedule_group_t
> group,
>  static int schedule_group_thrmask(odp_schedule_group_t group,
>                                   odp_thrmask_t *thrmask)
>  {
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>
>         if (group < 0 || group >= NUM_GROUP)
>                 return -1;
> @@ -765,7 +785,7 @@ static int schedule_group_thrmask(odp_schedule_group_t
> group,
>  static int schedule_group_info(odp_schedule_group_t group,
>                                odp_schedule_group_info_t *info)
>  {
> -       sched_group_t *sched_group = &sched_global.sched_group;
> +       sched_group_t *sched_group = &sched_global->sched_group;
>
>         if (group < 0 || group >= NUM_GROUP)
>                 return -1;
> --
> 2.7.4
>
>

Reply via email to