Re: [j-nsp] mx80 napt-44 with ms-mic on 13.2R5

2014-09-24 Thread Sergey Khalavchuk
Hello!

We had tested napt44 half a year ago, and napt44 configuration was
commiting successfuly but not functioning at all.
JTAC reported that napt44 is not supported on MIC, and created PR993320 for
us.
Based on commit error, this PR is fixed on your junos :)
We also been told that napt44 will probably appear at middle of next year
or about so.

--
wbr,
Sergey Khalavchuk


On Wed, Sep 24, 2014 at 8:47 AM, ryanL ryan.lan...@gmail.com wrote:

 has anyone been successful here? i'm getting the following error, even
 though juniper's docs seem to indicate this is supported on the ms-mic with
 13.2.

 my ref guides are:

 http://www.juniper.net/techpubs/en_US/junos13.2/information-products/topic-collections/config-guide-services/index.html?features-ms-mic.html

 http://www.juniper.net/techpubs/en_US/junos13.2/topics/example/nat-nat44-config-ms-mpc.html

 ry@iad1-er2# show | compare
 [edit]
 +  services {
 +  service-set SSET1 {
 +  nat-rules NAT-RULE1;
 +  interface-service {
 +  service-interface ms-0/2/0;
 +  }
 +  }
 +  nat {
 +  pool NP2 {
 +  address pub_space/28;
 +  port {
 +  automatic;
 +  }
 +  }
 +  rule NAT-RULE1 {
 +  match-direction input;
 +  term term-1 {
 +  from {
 +  source-address {
 +  10.0.0.0/8;
 +  }
 +  destination-address {
 +  10.0.0.0/8;
 +  }
 +  }
 +  then {
 +  no-translation;
 +  }
 +  }
 +  term term-2 {
 +  from {
 +  source-address {
 +  10.0.0.0/8;
 +  }
 +  }
 +  then {
 +  translated {
 +  source-pool NP2;
 +  translation-type {
 +  napt-44;
 +  }
 +  }
 +  }
 +  }
 +  }
 +  }
 +  }
 [edit interfaces]
 +   ms-0/2/0 {
 +   unit 0 {
 +   family inet;
 +   }
 +   }

 [edit]
 ry@iad1-er2# commit check
 [edit services]
   'service-set SSET1'
 translation type not supported on ms-interface
 error: configuration check-out failed

 [edit]
 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 https://puck.nether.net/mailman/listinfo/juniper-nsp




-- 
--
wbr
sergey khalavchuk
___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] mx80 napt-44 with ms-mic on 13.2R5

2014-09-24 Thread Alexander Arseniev

napt44 is most definitely is supported on MS-MIC
http://www.juniper.net/techpubs/en_US/junos13.2/topics/reference/general/nat-implementations-feature-comparison.html 


What is not supported is no-translation knob.
Please change Your config to (rough cut):
1/ delete term-1, and
2/ change term-2 to:

+  term term-2 {
+  from {
+  source-address {
+  10.0.0.0/8;
+  }
+  destination-address {
+  0.0.0.0/0;
+  10.0.0.0/8 except;
+  }
+  }
+  then {
+  translated {
+  source-pool NP2;
+  translation-type {
+  napt-44;
+  }

- then re-test and report back please.
Thanks
Alex

On 24/09/2014 06:47, ryanL wrote:

has anyone been successful here? i'm getting the following error, even
though juniper's docs seem to indicate this is supported on the ms-mic with
13.2.

my ref guides are:
http://www.juniper.net/techpubs/en_US/junos13.2/information-products/topic-collections/config-guide-services/index.html?features-ms-mic.html
http://www.juniper.net/techpubs/en_US/junos13.2/topics/example/nat-nat44-config-ms-mpc.html

ry@iad1-er2# show | compare
[edit]
+  services {
+  service-set SSET1 {
+  nat-rules NAT-RULE1;
+  interface-service {
+  service-interface ms-0/2/0;
+  }
+  }
+  nat {
+  pool NP2 {
+  address pub_space/28;
+  port {
+  automatic;
+  }
+  }
+  rule NAT-RULE1 {
+  match-direction input;
+  term term-1 {
+  from {
+  source-address {
+  10.0.0.0/8;
+  }
+  destination-address {
+  10.0.0.0/8;
+  }
+  }
+  then {
+  no-translation;
+  }
+  }
+  term term-2 {
+  from {
+  source-address {
+  10.0.0.0/8;
+  }
+  }
+  then {
+  translated {
+  source-pool NP2;
+  translation-type {
+  napt-44;
+  }
+  }
+  }
+  }
+  }
+  }
+  }
[edit interfaces]
+   ms-0/2/0 {
+   unit 0 {
+   family inet;
+   }
+   }

[edit]
ry@iad1-er2# commit check
[edit services]
   'service-set SSET1'
 translation type not supported on ms-interface
error: configuration check-out failed

[edit]
___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] pfe-x/x/x and pfh-a/a/a

2014-09-24 Thread junos fordummies
Hi Tom,

Did you manage to get an answer for this. I am also looking for an
explanation to interfaces that have been captured as part of a SNMP walk on
our MX.

Can anyone shed any light on what a pfe and pfh interface is ? Is there a
doc with all interface types that describe what these are ? I have searched
and couldn't find a satisfactory document.

Thanks

JFD

On 21 November 2011 17:22, Thomas Eichhorn t...@te3networks.de wrote:

 Dear all,

 I just saw some interfaces named
 pfe- and pfh- on a trio based
 MX960:

 teichhorn@R10-et show interfaces pfh-7/0/0
 Physical interface: pfh-7/0/0, Enabled, Physical link is Up
   Interface index: 135, SNMP ifIndex: 549
   Speed: 800mbps
   Device flags   : Present Running
   Link flags : None
   Last flapped   : Never
 Input packets : 0
 Output packets: 0

   Logical interface pfh-7/0/0.16383 (Index 328) (SNMP ifIndex 552)
 Flags: SNMP-Traps Encapsulation: ENET2
 Bandwidth: 0
 Input packets : 0
 Output packets: 0
 Protocol inet, MTU: Unlimited

 teichhorn@R10-et show interfaces pfe-7/0/0
 Physical interface: pfe-7/0/0, Enabled, Physical link is Up
   Interface index: 136, SNMP ifIndex: 550
   Speed: 800mbps
   Device flags   : Present Running
   Link flags : None
   Last flapped   : Never
 Input packets : 0
 Output packets: 0

   Logical interface pfe-7/0/0.16383 (Index 327) (SNMP ifIndex 553)
 Flags: SNMP-Traps Encapsulation: ENET2
 Bandwidth: 0
 Input packets : 0
 Output packets: 0
 Protocol inet, MTU: Unlimited
 Protocol inet6, MTU: Unlimited

 My best guess would be that pfe-x/x/x has something to do
 with the pfe, but pfh is completely unclear...

 Has someone here a little bit more information about the purpose
 of these ifaces?

 Thanks,
 Tom

 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 https://puck.nether.net/mailman/listinfo/juniper-nsp

___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] pfe-x/x/x and pfh-a/a/a

2014-09-24 Thread Edward Dore
Take a look at http://kb.juniper.net/InfoCenter/index?page=contentid=KB23578 
for a brief description of the pfh interfaces.

Edward Dore 
Freethought Internet 

On 24 Sep 2014, at 10:48, junos fordummies junos.fordumm...@googlemail.com 
wrote:

 Hi Tom,
 
 Did you manage to get an answer for this. I am also looking for an
 explanation to interfaces that have been captured as part of a SNMP walk on
 our MX.
 
 Can anyone shed any light on what a pfe and pfh interface is ? Is there a
 doc with all interface types that describe what these are ? I have searched
 and couldn't find a satisfactory document.
 
 Thanks
 
 JFD
 
 On 21 November 2011 17:22, Thomas Eichhorn t...@te3networks.de wrote:
 
 Dear all,
 
 I just saw some interfaces named
 pfe- and pfh- on a trio based
 MX960:
 
 teichhorn@R10-et show interfaces pfh-7/0/0
 Physical interface: pfh-7/0/0, Enabled, Physical link is Up
  Interface index: 135, SNMP ifIndex: 549
  Speed: 800mbps
  Device flags   : Present Running
  Link flags : None
  Last flapped   : Never
Input packets : 0
Output packets: 0
 
  Logical interface pfh-7/0/0.16383 (Index 328) (SNMP ifIndex 552)
Flags: SNMP-Traps Encapsulation: ENET2
Bandwidth: 0
Input packets : 0
Output packets: 0
Protocol inet, MTU: Unlimited
 
 teichhorn@R10-et show interfaces pfe-7/0/0
 Physical interface: pfe-7/0/0, Enabled, Physical link is Up
  Interface index: 136, SNMP ifIndex: 550
  Speed: 800mbps
  Device flags   : Present Running
  Link flags : None
  Last flapped   : Never
Input packets : 0
Output packets: 0
 
  Logical interface pfe-7/0/0.16383 (Index 327) (SNMP ifIndex 553)
Flags: SNMP-Traps Encapsulation: ENET2
Bandwidth: 0
Input packets : 0
Output packets: 0
Protocol inet, MTU: Unlimited
Protocol inet6, MTU: Unlimited
 
 My best guess would be that pfe-x/x/x has something to do
 with the pfe, but pfh is completely unclear...
 
 Has someone here a little bit more information about the purpose
 of these ifaces?
 
 Thanks,
 Tom
 
 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 https://puck.nether.net/mailman/listinfo/juniper-nsp
 
 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 https://puck.nether.net/mailman/listinfo/juniper-nsp

___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] pfe-x/x/x and pfh-a/a/a

2014-09-24 Thread junos fordummies
thanks Ed !!

On 24 September 2014 10:59, Edward Dore 
edward.d...@freethought-internet.co.uk wrote:

 Take a look at
 http://kb.juniper.net/InfoCenter/index?page=contentid=KB23578 for a
 brief description of the pfh interfaces.

 Edward Dore
 Freethought Internet

 On 24 Sep 2014, at 10:48, junos fordummies 
 junos.fordumm...@googlemail.com wrote:

 Hi Tom,

 Did you manage to get an answer for this. I am also looking for an
 explanation to interfaces that have been captured as part of a SNMP walk on
 our MX.

 Can anyone shed any light on what a pfe and pfh interface is ? Is there a
 doc with all interface types that describe what these are ? I have searched
 and couldn't find a satisfactory document.

 Thanks

 JFD

 On 21 November 2011 17:22, Thomas Eichhorn t...@te3networks.de wrote:

 Dear all,

 I just saw some interfaces named
 pfe- and pfh- on a trio based
 MX960:

 teichhorn@R10-et show interfaces pfh-7/0/0
 Physical interface: pfh-7/0/0, Enabled, Physical link is Up
  Interface index: 135, SNMP ifIndex: 549
  Speed: 800mbps
  Device flags   : Present Running
  Link flags : None
  Last flapped   : Never
Input packets : 0
Output packets: 0

  Logical interface pfh-7/0/0.16383 (Index 328) (SNMP ifIndex 552)
Flags: SNMP-Traps Encapsulation: ENET2
Bandwidth: 0
Input packets : 0
Output packets: 0
Protocol inet, MTU: Unlimited

 teichhorn@R10-et show interfaces pfe-7/0/0
 Physical interface: pfe-7/0/0, Enabled, Physical link is Up
  Interface index: 136, SNMP ifIndex: 550
  Speed: 800mbps
  Device flags   : Present Running
  Link flags : None
  Last flapped   : Never
Input packets : 0
Output packets: 0

  Logical interface pfe-7/0/0.16383 (Index 327) (SNMP ifIndex 553)
Flags: SNMP-Traps Encapsulation: ENET2
Bandwidth: 0
Input packets : 0
Output packets: 0
Protocol inet, MTU: Unlimited
Protocol inet6, MTU: Unlimited

 My best guess would be that pfe-x/x/x has something to do
 with the pfe, but pfh is completely unclear...

 Has someone here a little bit more information about the purpose
 of these ifaces?

 Thanks,
 Tom

 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 https://puck.nether.net/mailman/listinfo/juniper-nsp

 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 https://puck.nether.net/mailman/listinfo/juniper-nsp



___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] pfe-x/x/x and pfh-a/a/a

2014-09-24 Thread aqeel ahmed via juniper-nsp

Hello Mates
fpc/pic/port.16383 is an internally generated, nonconfigurable interface for 
router control traffic.
--
Sent from myMail app for Android
Wednesday, 24 September 2014, 00:48PM +03:00 from junos fordummies 
junos.fordumm...@googlemail.com:
Hi Tom,
Did you manage to get an answer for this. I am also looking for an
explanation to interfaces that have been captured as part of a SNMP walk on
our MX.
Can anyone shed any light on what a pfe and pfh interface is ? Is there a
doc with all interface types that describe what these are ? I have searched
and couldn't find a satisfactory document.
Thanks
JFD
On 21 November 2011 17:22, Thomas Eichhorn  t...@te3networks.de  wrote:
 Dear all,

 I just saw some interfaces named
 pfe- and pfh- on a trio based
 MX960:

 teichhorn@R10-et show interfaces pfh-7/0/0
 Physical interface: pfh-7/0/0, Enabled, Physical link is Up
   Interface index: 135, SNMP ifIndex: 549
   Speed: 800mbps
   Device flags   : Present Running
   Link flags : None
   Last flapped   : Never
 Input packets : 0
 Output packets: 0

   Logical interface pfh-7/0/0.16383 (Index 328) (SNMP ifIndex 552)
 Flags: SNMP-Traps Encapsulation: ENET2
 Bandwidth: 0
 Input packets : 0
 Output packets: 0
 Protocol inet, MTU: Unlimited

 teichhorn@R10-et show interfaces pfe-7/0/0
 Physical interface: pfe-7/0/0, Enabled, Physical link is Up
   Interface index: 136, SNMP ifIndex: 550
   Speed: 800mbps
   Device flags   : Present Running
   Link flags : None
   Last flapped   : Never
 Input packets : 0
 Output packets: 0

   Logical interface pfe-7/0/0.16383 (Index 327) (SNMP ifIndex 553)
 Flags: SNMP-Traps Encapsulation: ENET2
 Bandwidth: 0
 Input packets : 0
 Output packets: 0
 Protocol inet, MTU: Unlimited
 Protocol inet6, MTU: Unlimited

 My best guess would be that pfe-x/x/x has something to do
 with the pfe, but pfh is completely unclear...

 Has someone here a little bit more information about the purpose
 of these ifaces?

 Thanks,
 Tom

 ___
 juniper-nsp mailing list  juniper-nsp@puck.nether.net
  https://puck.nether.net/mailman/listinfo/juniper-nsp

___
juniper-nsp mailing list  juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp
___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp

Re: [j-nsp] MX80 Sampling - High CPU

2014-09-24 Thread Scott Granados
+1 here, definitely awaiting these releases.

On Sep 23, 2014, at 4:28 PM, Graham Brown juniper-...@grahambrown.info wrote:

 12.3R8 and 13.3R4 are due out anytime now with the fixes in place. I think
 there are many people waiting for these two releases...
 
 Cheers,
 
 Graham Brown
 Twitter - @mountainrescuer https://twitter.com/#!/mountainrescuer
 LinkedIn http://www.linkedin.com/in/grahamcbrown
 
 On 24 September 2014 03:18, Justin M. Streiner strei...@cluebyfour.org
 wrote:
 
 Sounds like you are running into bugs PR963060 or PR671136.
 
 This is supposed to be fixed in 12.3R8 which is supposed to be released
 very soon.
 
 We ran into this behavior on a pair of MX480s and had to disable sampling
 for the time being.
 
 jms
 
 
 On Tue, 23 Sep 2014, Ritz Rojas wrote:
 
 We have a few MX80s (MX80-48T) that we're looking to deploy in certain
 applications where they'll be taking full Internet tables (v4 and v6).  We
 also have a need to gather flow data on our routers, and have noticed an
 interesting trend in the lab.
 
 We are not using an MS-MIC currently.
 
 This test box is running 12.3R7.7 at the moment, but we've seen this same
 thing in 11.4 too.
 
 When set up with full Internet routes and sampling is enabled, each time a
 commit is made for any change at all, RPD and sampled take turns grinding
 the CPU up to 100%, for up to 5-10 minutes or more post-commit, and we see
 changes to BGP policy sometimes stall and take a decent amount of time (on
 the order of several minutes or more) to actually take effect.
 
 First RPD will climb up to almost 100% CPU utilization, chew it for a few
 minutes, then it'll go down and sampled will climb up to almost 100% for
 it's couple minutes turn and chew a bit.  Then sampled goes back down and
 RPD takes back over to 100% for a few more minutes.  Eventually it all
 finally calms back down and normalizes back to expected levels.
 
 Turn off sampling, and any CPU spikes post-commit are only on the order of
 seconds, not minutes, and any policy changes take effect pretty much
 immediately.
 
 We've seen this regardless of how flow is configured; we've configured
 flow
 with a simple config, as well as inline jflow, pretty much with the same
 results.  We're curious if anyone's had any of these same problems with
 jflow killing the CPU on MX80s (yeah, I know these PPC boxes are pretty
 weak sisters), and if there's any fix beyond the usual Doctor, it hurts
 when I do this, what should I do?.  Don't do that!.
 
 It's a nice feature, shame that using it seems to come with this heavy a
 price.
 
 As an aside, we also see a bit of a slowdown in the RIB/FIB
 learning/purging on BGP session turnup/reset, which we're well aware is a
 known issue with sampling enabled, so I won't be shocked if this is just
 how it is.  I'd love to be wrong.
 
 Here's our sampling config, quick and dirty, regular and inline jflow, in
 case we're missing something.
 
 Normal Sampling:
 
 router show configuration forwarding-options
 sampling {
   input {
   rate 8192;
   run-length 0;
   max-packets-per-second 2;
   }
   family inet {
   output {
   flow-server x.x.x.x {
   port x;
   version 5;
   }
   }
   }
 }
 
 router show configuration interfaces xe-0/0/0
 unit xxx {
   vlan-id xxx;
   family inet {
   sampling {
   input;
   output;
   }
 }
 
 
 Inline Jflow Sampling:
 
 router show configuration forwarding-options
 sampling {
   instance {
   BLAH-INSTANCE {
   input {
   rate 5000;
   }
   family inet {
   output {
   flow-server x.x.x.x {
   port ;
   autonomous-system-type origin;
   no-local-dump;
   version-ipfix {
   template {
   BLAH-TEMPLATE;
   }
   }
   }
   inline-jflow {
   source-address x.x.x.x;
   }
   }
   }
   }
   }
 }
 
 router show configuration chassis
 tfeb {
   slot 0 {
   sampling-instance BLAH-INSTANCE;
   }
 }
 
 
 router show configuration services
 flow-monitoring {
   version-ipfix {
   template BLAH-TEMPLATE {
   flow-active-timeout 10;
   flow-inactive-timeout 10;
   template-refresh-rate {
   packets 1;
   seconds 10;
   }
   option-refresh-rate {
   packets 1;
   seconds 10;
   }
   ipv4-template;
   }
   }
 }
 
 
 router show configuration interfaces xe-0/0/0
 unit xxx {
   vlan-id xxx;
   family inet {
   sampling {
   input;
   output;
   }
 }
 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 

Re: [j-nsp] mx80 napt-44 with ms-mic on 13.2R5

2014-09-24 Thread ryanL
thanks for the replies, folks. indeed it was the no-translation thing that
is hanging up the commit, and not the reported napt-44 statement. silly
defect.

i'm using this now:

rule NAT-RULE1 {
match-direction input;
term term-2 {
from {
source-address {
10.0.0.0/8;
}
destination-address {
0.0.0.0/0;
10.0.0.0/8 except;  # (good suggestion)
}
}
then {
translated {
source-pool NP2;
translation-type {
napt-44;

and that seems to commit ok.

however, implementing this on customer-facing interfaces broke the
customer, dropped BGP sessions, etc. my goal is to only nat traffic if it
is sourced from 10/8 and destined to anything other than 10/8. the NAT pool
is a static discard route exported in iBGP to ensure that this router
attracts return internet traffic in order to keep proper state.

what am i doing wrong? the documentation is quite poor for this module's
implementation, and sadly i don't have a lab to play with.

On Wed, Sep 24, 2014 at 2:13 AM, Alexander Arseniev arsen...@btinternet.com
 wrote:

 napt44 is most definitely is supported on MS-MIC
 http://www.juniper.net/techpubs/en_US/junos13.2/
 topics/reference/general/nat-implementations-feature-comparison.html
 What is not supported is no-translation knob.
 Please change Your config to (rough cut):
 1/ delete term-1, and
 2/ change term-2 to:

 +  term term-2 {
 +  from {
 +  source-address {
 +  10.0.0.0/8;
 +  }
 +  destination-address {
 +  0.0.0.0/0;
 +  10.0.0.0/8 except;
 +  }
 +  }
 +  then {
 +  translated {
 +  source-pool NP2;
 +  translation-type {
 +  napt-44;
 +  }

 - then re-test and report back please.
 Thanks
 Alex


 On 24/09/2014 06:47, ryanL wrote:

 has anyone been successful here? i'm getting the following error, even
 though juniper's docs seem to indicate this is supported on the ms-mic
 with
 13.2.

 my ref guides are:
 http://www.juniper.net/techpubs/en_US/junos13.2/
 information-products/topic-collections/config-guide-
 services/index.html?features-ms-mic.html
 http://www.juniper.net/techpubs/en_US/junos13.2/topics/example/nat-nat44-
 config-ms-mpc.html

 ry@iad1-er2# show | compare
 [edit]
 +  services {
 +  service-set SSET1 {
 +  nat-rules NAT-RULE1;
 +  interface-service {
 +  service-interface ms-0/2/0;
 +  }
 +  }
 +  nat {
 +  pool NP2 {
 +  address pub_space/28;
 +  port {
 +  automatic;
 +  }
 +  }
 +  rule NAT-RULE1 {
 +  match-direction input;
 +  term term-1 {
 +  from {
 +  source-address {
 +  10.0.0.0/8;
 +  }
 +  destination-address {
 +  10.0.0.0/8;
 +  }
 +  }
 +  then {
 +  no-translation;
 +  }
 +  }
 +  term term-2 {
 +  from {
 +  source-address {
 +  10.0.0.0/8;
 +  }
 +  }
 +  then {
 +  translated {
 +  source-pool NP2;
 +  translation-type {
 +  napt-44;
 +  }
 +  }
 +  }
 +  }
 +  }
 +  }
 +  }
 [edit interfaces]
 +   ms-0/2/0 {
 +   unit 0 {
 +   family inet;
 +   }
 +   }

 [edit]
 ry@iad1-er2# commit check
 [edit services]
'service-set SSET1'
  translation type not supported on ms-interface
 error: configuration check-out failed

 [edit]
 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 https://puck.nether.net/mailman/listinfo/juniper-nsp


 ___
 juniper-nsp mailing list juniper-nsp@puck.nether.net
 https://puck.nether.net/mailman/listinfo/juniper-nsp

___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] mx80 napt-44 with ms-mic on 13.2R5

2014-09-24 Thread Alexander Arseniev
I see You are using interface-style NAT - please exclude all control 
traffic in the service-filters from being diverted to MS-MIC, like below 
(rough cut, only BGP and BFD are excluded as an example):


set firewall family inet service-filter sf-in term 1 from protocol tcp
set firewall family inet service-filter sf-in term 1 from port 179
set firewall family inet service-filter sf-in term 1 then skip
set firewall family inet service-filter sf-in term 2 from protocol udp
set firewall family inet service-filter sf-in term 2 from port [ 3784 4784 ]
set firewall family inet service-filter sf-in term 2 then skip
set firewall family inet service-filter sf-in term 3 then service
set firewall family inet service-filter sf-out term 1 from protocol tcp
set firewall family inet service-filter sf-out term 1 from port 179
set firewall family inet service-filter sf-out term 1 then skip
set firewall family inet service-filter sf-out term 2 from protocol udp
set firewall family inet service-filter sf-out term 2 from port [ 3784 
4784 ]

set firewall family inet service-filter sf-out term 2 then skip
set firewall family inet service-filter sf-out term 3 then service
set interfaces ge-0/0/0.0 family inet service input service-set SSET1 
service-filter sf-in
set interfaces ge-0/0/0.0 family inet service output service-set SSET1 
service-filter sf-out


Alternatively, You can construct service-filters in such a way that only 
interested traffic is diverted to MS-MIC:


set firewall family inet service-filter sf-in term 1 from source-address 
10/8
set firewall family inet service-filter sf-in term 1 from 
destination-address 0/0
set firewall family inet service-filter sf-in term 1 from 
destination-address 10/8 except

set firewall family inet service-filter sf-in term 1 then service
set firewall family inet service-filter sf-in term 2 then skip

set firewall family inet service-filter sf-out term 1 from 
destination-address your nat pool route here

set firewall family inet service-filter sf-out term 1 then service
set firewall family inet service-filter sf-out term 2 then skip

set interfaces ge-0/0/0.0 family inet service input service-set SSET1 
service-filter sf-in
set interfaces ge-0/0/0.0 family inet service output service-set SSET1 
service-filter sf-out


HTH
Thanks
Alex

On 24/09/2014 17:37, ryanL wrote:
thanks for the replies, folks. indeed it was the no-translation thing 
that is hanging up the commit, and not the reported napt-44 statement. 
silly defect.


i'm using this now:

rule NAT-RULE1 {
match-direction input;
term term-2 {
from {
source-address {
10.0.0.0/8 http://10.0.0.0/8;
}
destination-address {
0.0.0.0/0 http://0.0.0.0/0;
10.0.0.0/8 http://10.0.0.0/8 except;  # (good suggestion)
}
}
then {
translated {
source-pool NP2;
translation-type {
napt-44;

and that seems to commit ok.

however, implementing this on customer-facing interfaces broke the 
customer, dropped BGP sessions, etc. my goal is to only nat traffic if 
it is sourced from 10/8 and destined to anything other than 10/8. the 
NAT pool is a static discard route exported in iBGP to ensure that 
this router attracts return internet traffic in order to keep proper 
state.


what am i doing wrong? the documentation is quite poor for this 
module's implementation, and sadly i don't have a lab to play with.


On Wed, Sep 24, 2014 at 2:13 AM, Alexander Arseniev 
arsen...@btinternet.com mailto:arsen...@btinternet.com wrote:


napt44 is most definitely is supported on MS-MIC

http://www.juniper.net/techpubs/en_US/junos13.2/topics/reference/general/nat-implementations-feature-comparison.html

What is not supported is no-translation knob.
Please change Your config to (rough cut):
1/ delete term-1, and
2/ change term-2 to:

+  term term-2 {
+  from {
+  source-address {
+ 10.0.0.0/8 http://10.0.0.0/8;
+  }
+  destination-address {
+ 0.0.0.0/0 http://0.0.0.0/0;
+ 10.0.0.0/8 http://10.0.0.0/8 except;
+  }
+  }
+  then {
+  translated {
+  source-pool NP2;
+  translation-type {
+  napt-44;
+  }

- then re-test and report back please.
Thanks
Alex


On 24/09/2014 06:47, ryanL wrote:

has anyone been successful here? i'm getting the following
error, even
though juniper's docs seem to indicate this is supported on
the ms-mic with
13.2.

my ref guides are:


Re: [j-nsp] mx80 napt-44 with ms-mic on 13.2R5

2014-09-24 Thread ryanL
alex, thank you. very helpful. of course i particularly like your
alternative, and that's what i thought i was accomplishing in the first
place. guess not!

can you explain the diff between `set firewall family inet service-filter`
vs the documentation example which references `set services
stateful-firewall` as part of the services set, and which is more
appropriate to use here in an ms-mic situation?

On Wed, Sep 24, 2014 at 10:00 AM, Alexander Arseniev 
arsen...@btinternet.com wrote:

  I see You are using interface-style NAT - please exclude all control
 traffic in the service-filters from being diverted to MS-MIC, like below
 (rough cut, only BGP and BFD are excluded as an example):

 set firewall family inet service-filter sf-in term 1 from protocol tcp
 set firewall family inet service-filter sf-in term 1 from port 179
 set firewall family inet service-filter sf-in term 1 then skip
 set firewall family inet service-filter sf-in term 2 from protocol udp
 set firewall family inet service-filter sf-in term 2 from port [ 3784 4784
 ]
 set firewall family inet service-filter sf-in term 2 then skip
 set firewall family inet service-filter sf-in term 3 then service
 set firewall family inet service-filter sf-out term 1 from protocol tcp
 set firewall family inet service-filter sf-out term 1 from port 179
 set firewall family inet service-filter sf-out term 1 then skip
 set firewall family inet service-filter sf-out term 2 from protocol udp
 set firewall family inet service-filter sf-out term 2 from port [ 3784
 4784 ]
 set firewall family inet service-filter sf-out term 2 then skip
 set firewall family inet service-filter sf-out term 3 then service
 set interfaces ge-0/0/0.0 family inet service input service-set SSET1
 service-filter sf-in
 set interfaces ge-0/0/0.0 family inet service output service-set SSET1
 service-filter sf-out

 Alternatively, You can construct service-filters in such a way that only
 interested traffic is diverted to MS-MIC:

 set firewall family inet service-filter sf-in term 1 from source-address
 10/8
 set firewall family inet service-filter sf-in term 1 from
 destination-address 0/0
 set firewall family inet service-filter sf-in term 1 from
 destination-address 10/8 except
 set firewall family inet service-filter sf-in term 1 then service
 set firewall family inet service-filter sf-in term 2 then skip

 set firewall family inet service-filter sf-out term 1 from
 destination-address your nat pool route here
 set firewall family inet service-filter sf-out term 1 then service
 set firewall family inet service-filter sf-out term 2 then skip

 set interfaces ge-0/0/0.0 family inet service input service-set SSET1
 service-filter sf-in
 set interfaces ge-0/0/0.0 family inet service output service-set SSET1
 service-filter sf-out

 HTH
 Thanks
 Alex

  On 24/09/2014 17:37, ryanL wrote:

 thanks for the replies, folks. indeed it was the no-translation thing that
 is hanging up the commit, and not the reported napt-44 statement. silly
 defect.

  i'm using this now:

  rule NAT-RULE1 {
 match-direction input;
 term term-2 {
 from {
 source-address {
 10.0.0.0/8;
 }
 destination-address {
 0.0.0.0/0;
 10.0.0.0/8 except;  # (good suggestion)
 }
 }
 then {
 translated {
 source-pool NP2;
 translation-type {
 napt-44;

  and that seems to commit ok.

  however, implementing this on customer-facing interfaces broke the
 customer, dropped BGP sessions, etc. my goal is to only nat traffic if it
 is sourced from 10/8 and destined to anything other than 10/8. the NAT pool
 is a static discard route exported in iBGP to ensure that this router
 attracts return internet traffic in order to keep proper state.

  what am i doing wrong? the documentation is quite poor for this module's
 implementation, and sadly i don't have a lab to play with.

 On Wed, Sep 24, 2014 at 2:13 AM, Alexander Arseniev 
 arsen...@btinternet.com wrote:

 napt44 is most definitely is supported on MS-MIC

 http://www.juniper.net/techpubs/en_US/junos13.2/topics/reference/general/nat-implementations-feature-comparison.html
 What is not supported is no-translation knob.
 Please change Your config to (rough cut):
 1/ delete term-1, and
 2/ change term-2 to:

 +  term term-2 {
 +  from {
 +  source-address {
 +  10.0.0.0/8;
 +  }
  +  destination-address {
 +  0.0.0.0/0;
 +  10.0.0.0/8 except;
 +  }
 +  }
 +  then {
 +  translated {
 +  source-pool NP2;
 +  translation-type {
 + 

Re: [j-nsp] Virtual Chassis RPD/BGP Rsync high CPU

2014-09-24 Thread Scott Harvanek
Okay so we traced this down to BGP Replication for NSR.  Looks like a 
bad attribute kills the replication process.  Other than blocking the 
received prefix is there a way to fix this:


Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]: Received malformed update from 
x
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Family inet-unicast, prefix 
5.56.168.0/21
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Malformed Attribute 
AGGREGATOR4(18) flag 0xc0 length 8.
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]: Total incoming malformed 
attributes from xx since last logging
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Received  1 malformed 
attribute AGGREGATOR4(18)


Mind you, the primary session with the peer stays up, this only kills 
the replication process...


Scott H.

On 9/18/14, 11:38 AM, Scott Harvanek wrote:
Has anyone had a issue with MX units in a VC where BGP rsync was 
consuming a boatload of CPU?


Master chassis shows:
Task   StartedUser Time  System Time Longest Run
BGP rsync 9650  10. 0.8  0.0
( BGP rsync is the only task with any user time during high user CPU 
for rpd )


now, that's only like 20% CPU on the master but on the slave it's 
90%  This seems to have happened when our total paths exceeded 2MM 
but does not seem to be a memory issue:


Dynamically allocated memory:  411009024  Maximum: 808517632
 Program data+BSS memory:5537792  Maximum: 5537792
  Page data overhead:1196032  Maximum: 1196032
 Page directory size: 212992  Maximum: 212992
  --
  Total bytes in use:  417955840 (12% of available memory)



___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] MX5 BNG licensing

2014-09-24 Thread Wojciech Janiszewski
Correct, unless you use other features that requires S-MX80-SSM-FP.

Regards,
Wojciech
___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] Virtual Chassis RPD/BGP Rsync high CPU

2014-09-24 Thread Alexander Arseniev
Have You tried drop-path-attributes 
http://kb.juniper.net/InfoCenter/index?page=contentid=JSA10491 ?

You can drop any attribute, not only 128 as in the KB.
Thanks
Alex

On 24/09/2014 18:38, Scott Harvanek wrote:
Okay so we traced this down to BGP Replication for NSR.  Looks like a 
bad attribute kills the replication process.  Other than blocking the 
received prefix is there a way to fix this:


Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]: Received malformed update from 
x
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Family inet-unicast, prefix 
5.56.168.0/21
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Malformed Attribute 
AGGREGATOR4(18) flag 0xc0 length 8.
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]: Total incoming malformed 
attributes from xx since last logging
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Received  1 malformed 
attribute AGGREGATOR4(18)


Mind you, the primary session with the peer stays up, this only kills 
the replication process...


Scott H.

On 9/18/14, 11:38 AM, Scott Harvanek wrote:
Has anyone had a issue with MX units in a VC where BGP rsync was 
consuming a boatload of CPU?


Master chassis shows:
Task   StartedUser Time  System Time Longest Run
BGP rsync 9650  10. 0.8  0.0
( BGP rsync is the only task with any user time during high user CPU 
for rpd )


now, that's only like 20% CPU on the master but on the slave it's 
90%  This seems to have happened when our total paths exceeded 
2MM but does not seem to be a memory issue:


Dynamically allocated memory:  411009024  Maximum: 808517632
 Program data+BSS memory:5537792  Maximum: 5537792
  Page data overhead:1196032  Maximum: 1196032
 Page directory size: 212992  Maximum: 212992
  --
  Total bytes in use:  417955840 (12% of available memory)



___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] mx80 napt-44 with ms-mic on 13.2R5

2014-09-24 Thread ryanL
very much appreciated, alex.

On Wed, Sep 24, 2014 at 11:01 AM, Alexander Arseniev 
arsen...@btinternet.com wrote:

  service-filter is a stateless firewall filter which has only 2 actions:
 either then service meaning divert to Service PIC (be it MS-PIC, MS-DPC
 NPU, MS-MIC etc) ; or then skip meaning do NOT divert to Service PIC,
 route the usual way.
 services stateful-firewall stanza is for defining stateful-firewall
 (SFW) services on Service PIC itself. This SFW service is less
 sophisticated than SRX SFW (filters basic attacks like Teardrop/Land but
 cannot filter more complicated attacks like Winnuke) but has more
 throughput.
 There are cases when You have to use both. I.e. to offer granulated SFW
 service to different customer subnets on same interface and/or spread the
 load between 1 Service PIC, You have to use both service-filter and
 services stateful-firewall stanzas.
 I suggest You read Junos Enterprise Routing book which has excellent
 primer on Juniper router services (interface-style vs nexthop style, SFW,
 NAT, also IPSec) - either 1st or 2nd edition is fine since basics did not
 change between the two.
 HTH
 Thanks
 Alex



 On 24/09/2014 18:13, ryanL wrote:

 alex, thank you. very helpful. of course i particularly like your
 alternative, and that's what i thought i was accomplishing in the first
 place. guess not!

 can you explain the diff between `set firewall family inet service-filter`
 vs the documentation example which references `set services
 stateful-firewall` as part of the services set, and which is more
 appropriate to use here in an ms-mic situation?

 On Wed, Sep 24, 2014 at 10:00 AM, Alexander Arseniev 
 arsen...@btinternet.com wrote:

  I see You are using interface-style NAT - please exclude all control
 traffic in the service-filters from being diverted to MS-MIC, like below
 (rough cut, only BGP and BFD are excluded as an example):

 set firewall family inet service-filter sf-in term 1 from protocol tcp
 set firewall family inet service-filter sf-in term 1 from port 179
 set firewall family inet service-filter sf-in term 1 then skip
 set firewall family inet service-filter sf-in term 2 from protocol udp
 set firewall family inet service-filter sf-in term 2 from port [ 3784
 4784 ]
 set firewall family inet service-filter sf-in term 2 then skip
 set firewall family inet service-filter sf-in term 3 then service
 set firewall family inet service-filter sf-out term 1 from protocol tcp
 set firewall family inet service-filter sf-out term 1 from port 179
 set firewall family inet service-filter sf-out term 1 then skip
 set firewall family inet service-filter sf-out term 2 from protocol udp
 set firewall family inet service-filter sf-out term 2 from port [ 3784
 4784 ]
 set firewall family inet service-filter sf-out term 2 then skip
 set firewall family inet service-filter sf-out term 3 then service
 set interfaces ge-0/0/0.0 family inet service input service-set SSET1
 service-filter sf-in
 set interfaces ge-0/0/0.0 family inet service output service-set SSET1
 service-filter sf-out

 Alternatively, You can construct service-filters in such a way that only
 interested traffic is diverted to MS-MIC:

 set firewall family inet service-filter sf-in term 1 from source-address
 10/8
 set firewall family inet service-filter sf-in term 1 from
 destination-address 0/0
 set firewall family inet service-filter sf-in term 1 from
 destination-address 10/8 except
 set firewall family inet service-filter sf-in term 1 then service
 set firewall family inet service-filter sf-in term 2 then skip

 set firewall family inet service-filter sf-out term 1 from
 destination-address your nat pool route here
 set firewall family inet service-filter sf-out term 1 then service
 set firewall family inet service-filter sf-out term 2 then skip

 set interfaces ge-0/0/0.0 family inet service input service-set SSET1
 service-filter sf-in
 set interfaces ge-0/0/0.0 family inet service output service-set SSET1
 service-filter sf-out

 HTH
 Thanks
 Alex

   On 24/09/2014 17:37, ryanL wrote:

 thanks for the replies, folks. indeed it was the no-translation thing
 that is hanging up the commit, and not the reported napt-44 statement.
 silly defect.

  i'm using this now:

  rule NAT-RULE1 {
 match-direction input;
 term term-2 {
 from {
 source-address {
 10.0.0.0/8;
 }
 destination-address {
 0.0.0.0/0;
 10.0.0.0/8 except;  # (good suggestion)
 }
 }
 then {
 translated {
 source-pool NP2;
 translation-type {
 napt-44;

  and that seems to commit ok.

  however, implementing this on customer-facing interfaces broke the
 customer, dropped BGP sessions, etc. my goal is to only nat traffic if
 it is sourced from 10/8 and destined to anything other 

Re: [j-nsp] Virtual Chassis RPD/BGP Rsync high CPU

2014-09-24 Thread Scott Harvanek
Well shoot, that's a great idea, looks like this command is hidden so 
I didn't even see it.  I assume AGGREGATOR4 is type code 18? I can't 
find anything official confirming that though?


Scott H.

On 9/24/14, 3:39 PM, Alexander Arseniev wrote:
Have You tried drop-path-attributes 
http://kb.juniper.net/InfoCenter/index?page=contentid=JSA10491 ?

You can drop any attribute, not only 128 as in the KB.
Thanks
Alex

On 24/09/2014 18:38, Scott Harvanek wrote:
Okay so we traced this down to BGP Replication for NSR.  Looks like a 
bad attribute kills the replication process.  Other than blocking the 
received prefix is there a way to fix this:


Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]: Received malformed update 
from x
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Family inet-unicast, prefix 
5.56.168.0/21
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Malformed Attribute 
AGGREGATOR4(18) flag 0xc0 length 8.
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]: Total incoming malformed 
attributes from xx since last logging
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Received  1 malformed 
attribute AGGREGATOR4(18)


Mind you, the primary session with the peer stays up, this only kills 
the replication process...


Scott H.

On 9/18/14, 11:38 AM, Scott Harvanek wrote:
Has anyone had a issue with MX units in a VC where BGP rsync was 
consuming a boatload of CPU?


Master chassis shows:
Task   StartedUser Time  System Time Longest 
Run

BGP rsync 9650  10. 0.8 0.0
( BGP rsync is the only task with any user time during high user CPU 
for rpd )


now, that's only like 20% CPU on the master but on the slave it's 
90%  This seems to have happened when our total paths exceeded 
2MM but does not seem to be a memory issue:


Dynamically allocated memory:  411009024  Maximum: 808517632
 Program data+BSS memory:5537792  Maximum: 5537792
  Page data overhead:1196032  Maximum: 1196032
 Page directory size: 212992  Maximum: 212992
  --
  Total bytes in use:  417955840 (12% of available memory)



___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] Virtual Chassis RPD/BGP Rsync high CPU

2014-09-24 Thread Scott Harvanek
Disregard, 18 is correct, looks like IETF/RFC4893 has this as  
AS4_AGGREGATOR not AGGREGATOR4.


Scott H.
Login Inc.

On 9/24/14, 5:11 PM, Scott Harvanek wrote:
Well shoot, that's a great idea, looks like this command is hidden 
so I didn't even see it.  I assume AGGREGATOR4 is type code 18? I 
can't find anything official confirming that though?


Scott H.

On 9/24/14, 3:39 PM, Alexander Arseniev wrote:
Have You tried drop-path-attributes 
http://kb.juniper.net/InfoCenter/index?page=contentid=JSA10491 ?

You can drop any attribute, not only 128 as in the KB.
Thanks
Alex

On 24/09/2014 18:38, Scott Harvanek wrote:
Okay so we traced this down to BGP Replication for NSR.  Looks like 
a bad attribute kills the replication process.  Other than blocking 
the received prefix is there a way to fix this:


Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]: Received malformed update 
from x
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Family inet-unicast, 
prefix 5.56.168.0/21
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Malformed Attribute 
AGGREGATOR4(18) flag 0xc0 length 8.
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]: Total incoming malformed 
attributes from xx since last logging
Sep 24 17:31:06  TUS-2-VC-1 rpd[48424]:   Received  1 malformed 
attribute AGGREGATOR4(18)


Mind you, the primary session with the peer stays up, this only 
kills the replication process...


Scott H.

On 9/18/14, 11:38 AM, Scott Harvanek wrote:
Has anyone had a issue with MX units in a VC where BGP rsync was 
consuming a boatload of CPU?


Master chassis shows:
Task   StartedUser Time  System Time 
Longest Run

BGP rsync 9650  10. 0.8 0.0
( BGP rsync is the only task with any user time during high user 
CPU for rpd )


now, that's only like 20% CPU on the master but on the slave it's 
90%  This seems to have happened when our total paths exceeded 
2MM but does not seem to be a memory issue:


Dynamically allocated memory:  411009024  Maximum: 808517632
 Program data+BSS memory:5537792  Maximum: 5537792
  Page data overhead:1196032  Maximum: 1196032
 Page directory size: 212992  Maximum: 212992
  --
  Total bytes in use:  417955840 (12% of available memory)



___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


___
juniper-nsp mailing list juniper-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/juniper-nsp


Re: [j-nsp] T4000 power architecture

2014-09-24 Thread Sam Silvester
Hi Fred,

Thanks for the reply. Your suggestions from step 1 - 3.2 makes sense.

Having said that, rather than swapping the PEMs in step 4, I can just swap
PEM1 with a replacement (we have support on the chassis). The difficulty
for me is we have to get electricians onsite out of hours to do this -
which I'm trying to avoid if it's unlikely the PEM is actually at fault.
This would be the case if the PEM output is carried to the FPCs via a
common bus arrangement, at which point I would suggest there is no way the
PEM can be at fault considering all the other FPCs DO show power being
received from PEM1. If I can confirm (via JTAC or otherwise) what the power
architecture is, then we can save this step, bite the bullet and move on
with getting the chassis replaced without wasting time on replacing the
PEM, then having to get a second window booked with electricians etc to
swap out the whole chassis.

I suppose ultimately part of the problem here has been that my coworkers
and myself would have assumed back in July when we first lodged this issue
to JTAC that asking a relatively straightforward question (how is the
power distributed from the PEMs to the FPCs in a T4000?) that we would
have actually had a fairly clear answer without too much fuss.

To date, that hasn't happened - we can't work out why this is; is it that
nobody runs T4000s? Is it that nobody in Juniper has this information? Is
it 'commercially sensitive' enough that nobody wants to tell us?

The response from JTAC so far has basically been hand-waving and avoiding
the question. That doesn't give me much confidence in either the platform
or JTAC.

Cheers,

Sam



On Thu, Sep 25, 2014 at 2:33 AM, Fred Quan fq...@juniper.net wrote:

 Hi Sam,

 Maybe you can try the below steps to isolate the problem.

 1.  run show chassis environment pem to check which FPC has 0 Voltage
 and Current on PEM.
 ( Here, I assume that,  FPC 0 has 0 Voltage and Current on PEM 1 )
 2. check FPC 0 is Online state with Dual PEM power on.
 3. power off PEM #0 in slot 0, then check FPC 0 state.
 3.1 if FPC 0 is still Online state with single PEM #1 power on, then PEM
 #1 in slot 1 can provide power to FPC 0, it is a display problem in STEP 1.
 3.2 if FPC 0 is Offline state with single PEM #1 power on, then PEM #1 in
 slot 1 cannot provide power to FPC 0, then continue STEP 4.
 4. power off both PEMs and swap PEM #0 and PEM #1, so PEM #0 in slot 1 and
 PEM #1 in slot 0.
 5. power on PEM #1 in slot 0 only. It looks like the PEM #1 problem if FPC
 0 is Offline state.
 6. power on PEM #0 in slot 1 only. It looks like the Mid plane problem if
 FPC 0 is Offline state.


 Regards,

 -Fred


 
 From: juniper-nsp juniper-nsp-boun...@puck.nether.net on behalf of Sam
 Silvester sam.silves...@gmail.com
 Sent: Tuesday, September 23, 2014 9:40 PM
 To: juniper-nsp@puck.nether.net
 Subject: Re: [j-nsp] T4000 power architecture

 Bumping for good measure...

 So JTAC are suggesting it's either the PEM or the midplane, but I was
 hoping to get some more information so I can validate what I'm being told -
 which, currently, is very little. Basically the suggestion is replace the
 PEM, if that doesn't work replace the chassis. I'd like to avoid
 whack-a-mole if possible; for example, if I knew that all of the FPCs were
 fed via the PEM via a central bus on the midplane, then we don't have to
 worry about replacing the PEM and can go straight ahead with organising a
 chassis swap.

 I don't suppose somebody happens to have a spare T4000 chassis or PEM lying
 around and could take a photo of the connectors between the PEM and the
 midplane? I'm still trying to get an idea of how power is distributed from
 the PEM to the actual FPC slots.

 The exact same questions has been posed to JTAC multiple times, but no
 answer has been forthcoming. I really hope the nagging suspicion I have
 that nobody in Juniper knows this information is incorrect but in the
 meantime I'm posting here in the hope I can figure it out from some
 pictures.

 Google image search isn't helping me find any clear photos of the rear of
 the PEM or the midplane :/

 Sam


 On Thu, Sep 11, 2014 at 8:06 AM, Sam Silvester sam.silves...@gmail.com
 wrote:

  Hi Aqeel - thanks for the reply.
 
  Agree 100% - the problem is, we only seem to be getting power to FPC0
 from
  one PEM... have a look below (and also note that FPC1 is fine, as are the
  rest of the FPCs).
 
  The question is - are we looking at a PEM fault here, or a midplane
 fault?
  As per previous discussion, the FPC itself seems fine, as moving it to
  another slot resolves the issue. Putting another card into slot 0 yields
  the same result as below.
 
  PEM 0 status:
State  Online
Temperature32 degrees C / 89 degrees F
DC Input:  OK
Voltage(V)  Current(A)  Power(W)  Load(%)
INPUT 0   54.750   4.312  236