Hi,

I ran a quick test with 9.2R2.15 between two BGP peers and I see BGP metric (MED) changes take effect immediately.

tbo...@brooklyn> show configuration protocols bgp
group test {
    type internal;
    local-address 50.50.50.1;
    family inet {
        unicast;
    }
    family inet-vpn {
        unicast;
    }
    export static-export;
    ipsec-sa bgp-secure;
    multipath;
    neighbor 50.50.50.254;
}

tbo...@manhattan> show configuration protocols bgp
group test {
    type internal;
    local-address 50.50.50.254;
    import set-med;
    family inet {
        unicast;
    }
    family inet-vpn {
        unicast;
    }
    ipsec-sa bgp-secure;
    neighbor 50.50.50.1;
}

tbo...@manhattan> show configuration policy-options policy-statement set-med
term 1 {
    from metric 0;
    then {
        metric 30000;
    }
}
term local_pref {
    then {
        local-preference 110;
        accept;
    }
}
term default {
    then reject;
}


Now I will start with no import policy on manhattan.

Sending 3 routes I see this:

tbo...@manhattan# run show route protocol bgp

inet.0: 10 destinations, 13 routes (10 active, 0 holddown, 0 hidden)
+ = Active Route, - = Last Active, * = Both

60.60.60.1/32       [BGP/170] 00:17:08, MED 100, localpref 100
                      AS path: I
                    > to 50.50.50.1 via em0.0
60.60.60.2/32       [BGP/170] 00:17:07, MED 0, localpref 100
                      AS path: I
                    > to 50.50.50.1 via em0.0
60.60.60.3/32       [BGP/170] 00:00:06, MED 300, localpref 100
                      AS path: I
                    > to 50.50.50.1 via em0.0

So now we want to turn on the import policy on manhattan, commit and see what happens.

tbo...@manhattan# run show route protocol bgp

inet.0: 10 destinations, 13 routes (10 active, 0 holddown, 0 hidden)
+ = Active Route, - = Last Active, * = Both

60.60.60.1/32       [BGP/170] 00:18:58, MED 100, localpref 110
                      AS path: I
                    > to 50.50.50.1 via em0.0
60.60.60.2/32       [BGP/170] 00:18:57, MED 30000, localpref 110
                      AS path: I
                    > to 50.50.50.1 via em0.0
60.60.60.3/32       [BGP/170] 00:01:56, MED 300, localpref 110
                      AS path: I
                    > to 50.50.50.1 via em0.0


This worked instantly without needing to clear the BGP session.

If you turn on traceoptions on BGP you should see something like this .. which shows the new policy being evaluated and then route attributes changed:

Jul 14 17:23:55.050048 peer 50.50.50.1 (test): Need to reevaluate import policy Jul 14 17:23:55.052131 task_timer_uset: timer BGP RT Background_BGP Route statistics timer <Touched> set to interval 30 at 17:24:25 Jul 14 17:23:55.052141 bgp_rt_update_med_igp_init: Deleting MED IGP update timer Jul 14 17:23:55.052147 group group test type Internal: export eval flag set (vpn nlri) Jul 14 17:23:55.052151 50.50.50.1 (Internal AS 1): import eval flag set (config change)
Jul 14 17:23:55.052472 init bgp commit sync

Jul 14 17:23:55.052495 bgp_rib_notify: freddy.inet.0 Add - exists
Jul 14 17:23:55.052531 task_job_create_background: create prio 5 job BGP reconfig for task BGP.0.0.0.0+179 Jul 14 17:23:55.061915 background dispatch running job BGP reconfig for task BGP.0.0.0.0+179 Jul 14 17:23:55.062080 CHANGE 60.60.60.1/32 gw 50.50.50.1 BGP pref 170/-111 metric 100/0 <Int Ext> as 1 Jul 14 17:23:55.062106 CHANGE 60.60.60.2/32 gw 50.50.50.1 BGP pref 170/-111 metric 30000/0 <Int Ext> as 1 Jul 14 17:23:55.062177 CHANGE 60.60.60.3/32 gw 50.50.50.1 BGP pref 170/-111 metric 300/0 <Int Ext> as 1

Not sure why it was necessary to hard clear the BGP session; does the upstream peer support BGP refresh?


Kind regards,
Truman Boyes











On 13/07/2009, at 6:35 PM, Will Orton wrote:

I have 2 POPs each with a connection to a common upstream. The upstream
is sending me MEDs, but lots of routes have (missing or 0) MEDs and I
want to reset those to a fixed value so I can tweak them later.

So I have an import policy on each BGP session like so:

term setall-meds {
   from metric 0;
   then {
       metric 30000;
   }
}
term def {
   then {
       local-preference 110;
       accept;
   }
}
term rej {
   then reject;
}


I apply this on both routers and get, for example:

At POP A (M10i 9.3R1.7):
A Destination P Prf Metric 1 Metric 2 Next hop AS path * 64.152.0.0/13 B 170 110 0 >(TO POP B) 3356 I B 170 110 30000 >(UPSTREAM AT A) 3356 I

At POP B (M10 9.3R3.8):
A Destination P Prf Metric 1 Metric 2 Next hop AS path * 64.152.0.0/13 B 170 110 0 >(UPSTREAM AT B) 3356 I


So the M10 at POP B doesn't appear to be applying the import policy and
setting the MED to 30000. POP A as a result picks the route through B.
(Yes, I waited more than the 15 minutes for POP B's CPU to go back to
idle so the RE-333-768 churned through the whole table).

This resolved itself with a hard clear of the BGP session to the upstream at POP B. 'soft-intbound' clear at B didn't do it (other than pegging the
RE CPU for another 15 minutes).

Any ideas? JUNOS bug? Old/decrepit RE getting non-deterministic
with age? Do I really have to hard-clear the BGP session on the 'B'
router any time I change the import policy now? :/


-Will Orton
_______________________________________________
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


_______________________________________________
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp

Reply via email to