Hi,

> thank you for your effort in improving the TT code :)

I try to make the TT code a better place.  ;-)


> > +   if (hard_iface == primary_if)
> > +           tt_num_changes = tt_append_changes(bat_priv,
> > +                                              &hard_iface->packet_buff,
> > +                                              &hard_iface->packet_len,
> > +                                              BATMAN_OGM_HLEN);
> > +
> > +   if (tt_num_changes > 0)
> > +           batman_ogm_packet->tt_num_changes = tt_num_changes;
> > +   else
> > +           batman_ogm_packet->tt_num_changes = 0;
> 
> Do we really need this if-loop? Am I wrong or tt_num_changes can only be >=
> 0 ?

Right, strictly speaking it is not needed. However, tt_append_changes() is 
defined as returning int, hence the calling function can't rely on this 
assumption.


> > -int tt_changes_fill_buffer(struct bat_priv *bat_priv,
> > -                      unsigned char *buff, int buff_len)
> > +static void tt_realloc_packet_buff(unsigned char **packet_buff,
> > +                              int *packet_buff_len, int min_packet_len,
> > +                              int new_packet_len)
> > +{
> > +   unsigned char *new_buff;
> > +
> > +   new_buff = kmalloc(new_packet_len, GFP_ATOMIC);
> > +
> > +   /* keep old buffer if kmalloc should fail */
> > +   if (new_buff) {
> > +           memcpy(new_buff, *packet_buff, min_packet_len);
> > +           kfree(*packet_buff);
> > +           *packet_buff = new_buff;
> > +           *packet_buff_len = new_packet_len;
> > +   }
> 
> I took quite a while to understand what happens to packet_buff_len if
> kmalloc failed. Actually it correctly stores the "previous" buffer size, so
> the rest of the code will handle kmalloc failures the right way. :)

Actually, this part of the code did not change. Check realloc_packet_buffer() 
in send.c and you will find the same function.


> > +}
> > +
> > +static void tt_prepare_packet_buff(struct bat_priv *bat_priv,
> > +                              unsigned char **packet_buff,
> > +                              int *packet_buff_len, int min_packet_len)
> > +{
> > +   struct hard_iface *primary_if;
> > +   int req_len;
> > +
> > +   primary_if = primary_if_get_selected(bat_priv);
> > +
> > +   req_len = min_packet_len;
> > +   req_len += tt_len((uint8_t)atomic_read(&bat_priv->tt_local_changes));
> 
> This cast is also in the current code. But why not removing it? atomic_t is
> an int, the tt_len() argument too.

No idea why the cast is there. I'll remove it.  :-)


> > +
> > +   /* if we have too many changes for one packet don't send any
> > +    * and wait for the tt table request which will be fragmented */
> 
> please fix this comment. */ must be on a new line.

Ok, I'll fix it. Just a quick reminder that this is old code as well ..


> > +static int tt_changes_fill_buff(struct bat_priv *bat_priv,
> > +                           unsigned char **packet_buff,
> > +                           int *packet_buff_len, int min_packet_len)
> > 
> >  {
> > 
> > -   int count = 0, tot_changes = 0;
> > 
> >     struct tt_change_node *entry, *safe;
> > 
> > +   int count = 0, tot_changes = 0, new_len;
> > +   unsigned char *tt_buff;
> > +
> 
> As suggesting on IRC we should lock the "read and copy procedure".
> I'd call lock() here.
> 
> > +   tt_prepare_packet_buff(bat_priv, packet_buff,
> > +                          packet_buff_len, min_packet_len);
> > 
> > -   if (buff_len > 0)
> > -           tot_changes = buff_len / tt_len(1);
> > +   new_len = *packet_buff_len - min_packet_len;
> > 
> > 
> > 
> > +   tt_buff = *packet_buff + min_packet_len;
> > +
> > +   if (new_len > 0)
> > +           tot_changes = new_len / tt_len(1);
> > 
> >     spin_lock_bh(&bat_priv->tt_changes_list_lock);
> >     atomic_set(&bat_priv->tt_local_changes, 0);
> > 
> > @@ -290,7 +339,7 @@ int tt_changes_fill_buffer(struct bat_priv *bat_priv,
> > 
> >     list_for_each_entry_safe(entry, safe, &bat_priv->tt_changes_list,
> >     
> >                              list) {
> >             
> >             if (count < tot_changes) {
> > 
> > -                   memcpy(buff + tt_len(count),
> > +                   memcpy(tt_buff + tt_len(count),
> > 
> >                            &entry->change, sizeof(struct tt_change));
> >                     
> >                     count++;
> >             
> >             }
> 
> and I'd call unlock() after having copied everything to the tt_buff and
> emptied the changes list. Can we directly use
> bat_priv->tt_changes_list_lock ? It seems to be the case :)

I'd rather move the locking into a separate patch to make it easier to trace 
the change.


> > 
> >     /* all the reset entries have now to be effectively counted as local
> >     
> >      * entries */
> >     
> >     atomic_add(changed_num, &bat_priv->num_local_tt);
> >     tt_local_purge_pending_clients(bat_priv);
> > 
> > +   bat_priv->tt_crc = tt_local_crc(bat_priv);
> > 
> >     /* Increment the TTVN only once per OGM interval */
> >     atomic_inc(&bat_priv->ttvn);
> >     bat_dbg(DBG_TT, bat_priv, "Local changes committed, updating to ttvn
> >     %u\n",
> >     
> >             (uint8_t)atomic_read(&bat_priv->ttvn));
> >     
> >     bat_priv->tt_poss_change = false;
> > 
> > +
> > +   /* reset the sending counter */
> > +   atomic_set(&bat_priv->tt_ogm_append_cnt, TT_OGM_APPEND_MAX);
> > +
> > +   return tt_changes_fill_buff(bat_priv, packet_buff,
> > +                               packet_buff_len, packet_min_len);
> > +}
> 
> As you suggested on IRC, we may want to envelop this function with a
> lock/unlock to force exclusive access to the local table and to the event
> list.
>
> We should apply the same lock in tt_local_add()/del() I think.


Why do want to lock tt_changes_fill_buff() and tt_commit_changes() separately? 
We should already lock in tt_commit_changes() because the entire commit has to 
be an atomic operation. Several of the function calls in tt_commit_changes() 
depend on the fact that no client is purged or added while these functions 
run.

Thanks for your comments!

Cheers,
Marek

Reply via email to