Re: [j-nsp] Juniper M120 - SSD Replacement
Hi, I'm guessing juniper's are over priced? I have yet to yank mine from our MX240 demo to see if a generic one would suffice... - Alain Hebertaheb...@pubnix.net PubNIX Inc. 50 boul. St-Charles P.O. Box 26770 Beaconsfield, Quebec H9W 6G7 Tel: 514-990-5911 http://www.pubnix.netFax: 514-990-9443 On 05/01/18 12:05, Juan C. Crespo R. wrote: Hello Guys Could you please tell me a good SSD replacement for this Routing Engine (RE-A-2000)?? thanks ___ juniper-nsp mailing list juniper-nsp@puck.nether.net https://puck.nether.net/mailman/listinfo/juniper-nsp ___ juniper-nsp mailing list juniper-nsp@puck.nether.net https://puck.nether.net/mailman/listinfo/juniper-nsp
Re: [j-nsp] migration from cisco VRF+Vrrp to the juniper ACX
does anyone have an idea why it does not work on Acx( vrf+ vrrp). Br ap Op di 24 apr. 2018 om 10:20 schreef A. Camci> > Hi Guys, >> >> we are migration the Cisco CISCO7606-S to the acx5096. >> But we have 1 customer with VRF and VRRP on the same port. >> >> after migration to the ACX has customer no connection from the VRF. >> if we switch back to cisco, everything works fine. >> >> this is a full redundant vrf. >> other side is still cisco and all locations are now running on the backup >> vrf. >> >> if we lower the priotry of the vrrp on the backup vrf we see that the >> primary location becomes master. so the vrrp does work. after switching >> the vrrp has customer still one way traffic from the acx. maybe vrf+vrrp >> doesnt work on a ACX. >> >> see below for the config. >> >> CISCO config >> >> vlan 3021 >> mtu 1600 >> ! >> interface Vlan3021 >> mtu 1600 >> ip vrf forwarding CUST_APPIE >> ip address 172.21.1.251 255.255.255.0 >> vrrp 1 description CUST_APPIE-DC-centraal-pri >> vrrp 1 ip 172.21.1.250 >> vrrp 1 preempt delay minimum 10 >> vrrp 1 priority 110 >> ! >> >> interface Te3/4 >> switchport trunk allowed vlan add 3021 >> >> ip vrf CUST_APPIE >> rd 10.31.0.61:10006 >> route-target export 65001:10006 >> route-target import 65001:10006 >> >> >> router bgp 65001 >> address-family ipv4 vrf CUST_APPIE >> no synchronization >> redistribute static >> redistribute connected >> default-information originate >> exit-address-family >> >> ip route vrf CUST_APPIE 0.0.0.0 0.0.0.0 172.21.1.1 >> >> >> JUNIPER CONFIG >> ACX Model: acx5096_ Junos: 15.1X54-D61.6 >> >> set interfaces xe-0/0/88 description "*** CUST_APPIE***" >> set interfaces xe-0/0/88 flexible-vlan-tagging >> set interfaces xe-0/0/88 speed 10g >> set interfaces xe-0/0/88 mtu 1622 >> set interfaces xe-0/0/88 encapsulation flexible-ethernet-services >> set interfaces xe-0/0/88 ether-options no-auto-negotiation >> >> set interfaces xe-0/0/88 unit 3021 vlan-id 3021 >> set interfaces xe-0/0/88 unit 3021 family inet address 172.21.1.251/24 >> vrrp-group 1 virtual-address 172.21.1.250 >> set interfaces xe-0/0/88 unit 3021 family inet address 172.21.1.251/24 >> vrrp-group 1 priority 110 >> set interfaces xe-0/0/88 unit 3021 family inet address 172.21.1.251/24 >> vrrp-group 1 preempt hold-time 10 >> set interfaces xe-0/0/88 unit 3021 family inet address 172.21.1.251/24 >> vrrp-group 1 accept-data >> >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-export term 1 >> from protocol direct >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-export term 1 >> from protocol static >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-export term 1 >> then accept >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-import term 1 >> from protocol bgp >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-import term 1 >> from route-filter 0.0.0.0/0 exact >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-import term 1 >> then local-preference 150 >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-import term 1 >> then accept >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-import term 2 >> from protocol direct >> set policy-options policy-statement ipvpn-CUST_APPIE-ebgp-import term 2 >> then accept >> >> set routing-instances CUST_APPIE instance-type vrf >> set routing-instances CUST_APPIE interface xe-0/0/88.3021 >> set routing-instances CUST_APPIE route-distinguisher 10.32.0.43:10006 >> set routing-instances CUST_APPIE vrf-target import target:65001:10006 >> set routing-instances CUST_APPIE vrf-target export target:65001:10006 >> set routing-instances CUST_APPIE vrf-table-label >> >> set routing-instances CUST_APPIE routing-options static route 0.0.0.0/0 >> next-hop 172.21.1.1 >> >> set routing-instances CUST_APPIE forwarding-options dhcp-relay >> server-group CUST_APPIE 172.21.1.1 >> set routing-instances CUST_APPIE forwarding-options dhcp-relay >> active-server-group CUST_APPIE >> set routing-instances CUST_APPIE forwarding-options dhcp-relay group >> CUST_APPIE interface xe-0/0/88.3021 >> set routing-instances CUST_APPIE protocols bgp group ebgp-CUST_APPIE >> import ipvpn-CUST_APPIE-ebgp-import >> set routing-instances CUST_APPIE protocols bgp group ebgp-CUST_APPIE >> export ipvpn-CUST_APPIE-ebgp-export >> >> set firewall family inet filter re-protect-v4 term accept-customer-vrrp >> from protocol vrrp >> set firewall family inet filter re-protect-v4 term accept-customer-vrrp >> then count accept-vrrp-customer >> set firewall family inet filter re-protect-v4 term accept-customer-vrrp >> then accept >> set firewall family inet filter routing-engine-traffic term mark-vrrp >> from protocol vrrp >> set firewall family inet filter routing-engine-traffic term mark-vrrp >> then count mark-vrrp >> set firewall family inet filter routing-engine-traffic term mark-vrrp >> then forwarding-class NC1 >> set firewall family inet filter routing-engine-traffic
[j-nsp] Juniper M120 - SSD Replacement
Hello Guys Could you please tell me a good SSD replacement for this Routing Engine (RE-A-2000)?? thanks ___ juniper-nsp mailing list juniper-nsp@puck.nether.net https://puck.nether.net/mailman/listinfo/juniper-nsp
Re: [j-nsp] MX104 and NetFlow - Any horror story to share?
❦ 1 mai 2018 14:30 GMT, Michael Hare: > chassis { > afeb { > slot 0 { > inline-services { > flow-table-size { > ipv4-flow-table-size 7; > ipv6-flow-table-size 7; > } > } > } > } > } On 15.1R6, I am using this without any issue: afeb { slot 0 { inline-services { flow-table-size { ipv4-flow-table-size 10; ipv6-flow-table-size 5; } } } } -- Don't sacrifice clarity for small gains in "efficiency". - The Elements of Programming Style (Kernighan & Plauger) ___ juniper-nsp mailing list juniper-nsp@puck.nether.net https://puck.nether.net/mailman/listinfo/juniper-nsp
Re: [j-nsp] MX104 and NetFlow - Any horror story to share?
Alain, Do you want to collect IPv6? You are probably passed 14.X code on MX104 but I observed that I was unable to change the ipv6-flow-table-size at all (including after reboot). I was able to set flow-table-size in 16.X but my load average on 16.X on MX104 is pretty terrible; seems like I got all of the performance penalty of threading in 16.X without an additional core unlocked on the MX104 RE. Since 14.X is near EOL I didn't harass JTAC. Thanks and a nod to Olivier, I hadn't seen "flex-flow-sizing" before, seems like I really wanted that, not the explicit flow-table-size commands. abbreviated code example below. chassis { afeb { slot 0 { inline-services { flow-table-size { ipv4-flow-table-size 7; ipv6-flow-table-size 7; } } } } } -Michael >>-Original Message- >>From: juniper-nsp [mailto:juniper-nsp-boun...@puck.nether.net] On Behalf >>Of Alain Hebert >>Sent: Tuesday, May 01, 2018 8:23 AM >>To: juniper-nsp@puck.nether.net >>Subject: Re: [j-nsp] MX104 and NetFlow - Any horror story to share? >> >> Yeah I had the feeling I would break those MX's. >> >> At this point it is worth it to rebuilt our vMX lab to test the >>IPFIX variant... >> >> Thanks for the input. >> >> >> As for routing we have a pretty good mix of T1/T2 providers and we >>rarely drop sessions so it is providing a pretty good uptime... And >>that's why we got a pair of MX960 coming down anytime this year. >> >> >> PS: Unrelated quote - Yeah fat fingers sorry list. >> >>- >>Alain Hebertaheb...@pubnix.net >>PubNIX Inc. >>50 boul. St-Charles >>P.O. Box 26770 Beaconsfield, Quebec H9W 6G7 >>Tel: 514-990-5911 http://www.pubnix.netFax: 514-990-9443 >> >>On 04/30/18 19:41, Olivier Benghozi wrote: >>> Hi Alain, >>> >>> While you seem to already be kind of suicidal (5 full tables peers on an >>MX104), on an MX you must not use netflow v9 (CPU based) but use inline >>IPFIX (Trio / PFE based). >>> I suppose that Netflow-v9 on an MX104 could be quickly an interesting >>horror story with real traffic due to its ridiculously slow CPU, by the way. >>> With inline IPFIX it should just take some more RAM, and FIB update could >>be a bit slower. >>> >>> By the way on MX104 you don't configure «fpc» (bigger MXs) of «tfeb» >>(MX80) in chassis hierarchy, but «afeb», so you can remove your fpc line and >>fix your tfeb line. >>> >>> So you'll need something like that in services, instead of version9: >>> set services flow-monitoring version-ipfix template ipv4 template-refresh- >>rate >>> set services flow-monitoring version-ipfix template ipv4 option-refresh- >>rate >>> set services flow-monitoring version-ipfix template ipv4 ipv4-template >>> >>> And these ones too, to allocate some memory for the flows in the Trio and >>to define how it will speaks with the collector: >>> set chassis afeb slot 0 inline-services flex-flow-sizing >>> set forwarding-options sampling instance NETFLOW-SI family inet output >>inline-jflow source-address a.b.c.d >>> >>> Of course you'll remove the line with «output flow-server source >>». >>> >>> >>> >>> I don't see why you quoted the mail from Brijesh Patel about the Routing >>licences, by the way :P >>> >>> >>> Olivier >>> On 30 apr. 2018 at 21:34, Alain Hebertwrote : Anyone has any horror stories with something similar to what we're about >>to do? We're planning to turn up the following Netflow config (see below) on >>our MX104s (while we wait for our new MX960 =D), it worked well with >>everything else (SRX mostly), the "*s**et chassis"* are making us wonder >>how high would be the possibility to render those system unstable, at short >>and long term. Thanks again for your time. PS: We're using Elastiflow, and its working great for our needs atm. -- A bit of context Model: mx104 Junos: 16.1R4-S1.3 They're routing about 20Gbps atm, with 5 full tables peers, ~0.20 load >>average, and 700MB mem free. -- The Netflow config *set chassis tfeb0 slot 0 sampling-instance NETFLOW-SI* *set chassis fpc 1 sampling-instance NETFLOW-SI* set services flow-monitoring version9 template FM-V9 option-refresh- >>rate seconds 25 set services flow-monitoring version9 template FM-V9 template-refresh- >>rate seconds 15 set services flow-monitoring version9 template FM-V9 ipv4-template set forwarding-options sampling instance NETFLOW-SI input rate 1 run- >>length 0 set forwarding-options sampling instance NETFLOW-SI family inet output >>flow-server port 2055 set forwarding-options sampling instance NETFLOW-SI family inet
Re: [j-nsp] Difference between MPC4E-3D-32XGE-RB and MPC4E-3D-32XGE-SFPP ?
Can’t remember the exact numbers but the non-RB card is targeted at MPLS core applications where it’s just high density label switching. Won’t take a full routing table and has reduced L3VPN numbers. Ask your AM/SE for the specifics. Sent from my iPhone > On 30 Apr 2018, at 10:34 am, Brijesh Patelwrote: > > Hello Members, > > Any idea what is Difference between MPC4E-3D-32XGE-RB and > MPC4E-3D-32XGE-SFPP ? > > Juniper PDf says : > > MPC4E-3D-32XGE-SFPP 32x10GbE, full scale L2/L2.5 and *reduced scale L3 > features* > and > MPC4E-3D-32XGE-RB 32XGbE SFPP ports, full scale L2/L2.5, > * L3 and L3VPN features* > > now question is *what is reduced scale L3 featurs and L3vpn features ?* > > *Many Thanks,* > > *Brijesh Patel* > ___ > juniper-nsp mailing list juniper-nsp@puck.nether.net > https://puck.nether.net/mailman/listinfo/juniper-nsp ___ juniper-nsp mailing list juniper-nsp@puck.nether.net https://puck.nether.net/mailman/listinfo/juniper-nsp