On Thu, Apr 26, 2012 at 01:49:29PM +0800, Marek Lindner wrote: > > Hi, > > > thank you for your effort in improving the TT code :) > > I try to make the TT code a better place. ;-)
mh, I think more effort is required! :-D
>
>
> > > + if (hard_iface == primary_if)
> > > + tt_num_changes = tt_append_changes(bat_priv,
> > > + &hard_iface->packet_buff,
> > > + &hard_iface->packet_len,
> > > + BATMAN_OGM_HLEN);
> > > +
> > > + if (tt_num_changes > 0)
> > > + batman_ogm_packet->tt_num_changes = tt_num_changes;
> > > + else
> > > + batman_ogm_packet->tt_num_changes = 0;
> >
> > Do we really need this if-loop? Am I wrong or tt_num_changes can only be >=
> > 0 ?
>
> Right, strictly speaking it is not needed. However, tt_append_changes() is
> defined as returning int, hence the calling function can't rely on this
> assumption.
>
Ok. Then tt_append_changes() should be fixed. It's a really stupid thing,
but it would simplify this part of the code.
>
> > > -int tt_changes_fill_buffer(struct bat_priv *bat_priv,
> > > - unsigned char *buff, int buff_len)
> > > +static void tt_realloc_packet_buff(unsigned char **packet_buff,
> > > + int *packet_buff_len, int min_packet_len,
> > > + int new_packet_len)
> > > +{
> > > + unsigned char *new_buff;
> > > +
> > > + new_buff = kmalloc(new_packet_len, GFP_ATOMIC);
> > > +
> > > + /* keep old buffer if kmalloc should fail */
> > > + if (new_buff) {
> > > + memcpy(new_buff, *packet_buff, min_packet_len);
> > > + kfree(*packet_buff);
> > > + *packet_buff = new_buff;
> > > + *packet_buff_len = new_packet_len;
> > > + }
> >
> > I took quite a while to understand what happens to packet_buff_len if
> > kmalloc failed. Actually it correctly stores the "previous" buffer size, so
> > the rest of the code will handle kmalloc failures the right way. :)
>
> Actually, this part of the code did not change. Check realloc_packet_buffer()
> in send.c and you will find the same function.
>
Yeah, this luckily means that the current code is correct :-)
>
> > > +}
> > > +
> > > +static void tt_prepare_packet_buff(struct bat_priv *bat_priv,
> > > + unsigned char **packet_buff,
> > > + int *packet_buff_len, int min_packet_len)
> > > +{
> > > + struct hard_iface *primary_if;
> > > + int req_len;
> > > +
> > > + primary_if = primary_if_get_selected(bat_priv);
> > > +
> > > + req_len = min_packet_len;
> > > + req_len += tt_len((uint8_t)atomic_read(&bat_priv->tt_local_changes));
> >
> > This cast is also in the current code. But why not removing it? atomic_t is
> > an int, the tt_len() argument too.
>
> No idea why the cast is there. I'll remove it. :-)
>
>
> > > +
> > > + /* if we have too many changes for one packet don't send any
> > > + * and wait for the tt table request which will be fragmented */
> >
> > please fix this comment. */ must be on a new line.
>
> Ok, I'll fix it. Just a quick reminder that this is old code as well ..
Yep, I know, but since you are touching that comment it is better to fix it :)
>
>
> > > +static int tt_changes_fill_buff(struct bat_priv *bat_priv,
> > > + unsigned char **packet_buff,
> > > + int *packet_buff_len, int min_packet_len)
> > >
> > > {
> > >
> > > - int count = 0, tot_changes = 0;
> > >
> > > struct tt_change_node *entry, *safe;
> > >
> > > + int count = 0, tot_changes = 0, new_len;
> > > + unsigned char *tt_buff;
> > > +
> >
> > As suggesting on IRC we should lock the "read and copy procedure".
> > I'd call lock() here.
> >
> > > + tt_prepare_packet_buff(bat_priv, packet_buff,
> > > + packet_buff_len, min_packet_len);
> > >
> > > - if (buff_len > 0)
> > > - tot_changes = buff_len / tt_len(1);
> > > + new_len = *packet_buff_len - min_packet_len;
> > >
> > >
> > >
> > > + tt_buff = *packet_buff + min_packet_len;
> > > +
> > > + if (new_len > 0)
> > > + tot_changes = new_len / tt_len(1);
> > >
> > > spin_lock_bh(&bat_priv->tt_changes_list_lock);
> > > atomic_set(&bat_priv->tt_local_changes, 0);
> > >
> > > @@ -290,7 +339,7 @@ int tt_changes_fill_buffer(struct bat_priv *bat_priv,
> > >
> > > list_for_each_entry_safe(entry, safe, &bat_priv->tt_changes_list,
> > >
> > > list) {
> > >
> > > if (count < tot_changes) {
> > >
> > > - memcpy(buff + tt_len(count),
> > > + memcpy(tt_buff + tt_len(count),
> > >
> > > &entry->change, sizeof(struct tt_change));
> > >
> > > count++;
> > >
> > > }
> >
> > and I'd call unlock() after having copied everything to the tt_buff and
> > emptied the changes list. Can we directly use
> > bat_priv->tt_changes_list_lock ? It seems to be the case :)
>
> I'd rather move the locking into a separate patch to make it easier to trace
> the change.
I agree
>
>
> > >
> > > /* all the reset entries have now to be effectively counted as local
> > >
> > > * entries */
> > >
> > > atomic_add(changed_num, &bat_priv->num_local_tt);
> > > tt_local_purge_pending_clients(bat_priv);
> > >
> > > + bat_priv->tt_crc = tt_local_crc(bat_priv);
> > >
> > > /* Increment the TTVN only once per OGM interval */
> > > atomic_inc(&bat_priv->ttvn);
> > > bat_dbg(DBG_TT, bat_priv, "Local changes committed, updating to ttvn
> > > %u\n",
> > >
> > > (uint8_t)atomic_read(&bat_priv->ttvn));
> > >
> > > bat_priv->tt_poss_change = false;
> > >
> > > +
> > > + /* reset the sending counter */
> > > + atomic_set(&bat_priv->tt_ogm_append_cnt, TT_OGM_APPEND_MAX);
> > > +
> > > + return tt_changes_fill_buff(bat_priv, packet_buff,
> > > + packet_buff_len, packet_min_len);
> > > +}
> >
> > As you suggested on IRC, we may want to envelop this function with a
> > lock/unlock to force exclusive access to the local table and to the event
> > list.
> >
> > We should apply the same lock in tt_local_add()/del() I think.
>
>
> Why do want to lock tt_changes_fill_buff() and tt_commit_changes()
> separately?
> We should already lock in tt_commit_changes() because the entire commit has
> to
> be an atomic operation. Several of the function calls in tt_commit_changes()
> depend on the fact that no client is purged or added while these functions
> run.
Yeah, lock in tt_commit_changes() is definitely safer. I was trying to reduce
the critical zone, but I'd open race conditions in that way.
Thank you!
--
Antonio Quartulli
..each of us alone is worth nothing..
Ernesto "Che" Guevara
pgpWWXrAcNUWd.pgp
Description: PGP signature
