2. 2
LANswitch 3LS--New Product Achievement Award Finalist--ComNet Feb 1997
Communications Week
MAX Award 1996
IDG Reseaux & Telecoms,
Champion of All Categories,
May 1995
Data Communications
Tester’s Choice Winner
November 1996
Byte Magazine,
Best Overall Award,
July 1995
LAN Times,
“Best of Times”
Switching Shootout,
February 1995
Data Communications Magazine,
Top Performer,
February 1996
PC Digest Ratings Report,
Best Store-and-Forward Ethernet Switch,
May 1995
SNCI Bradner Test,
Top Performer,
February 1995
LANswitch Plus:LANswitch Plus:
Multiple award winning platformMultiple award winning platform
4. 4
LANswitchLANswitch PlusPlus--
Architectural OverviewArchitectural Overview
Completely distributed switch
architecture featuring:
• 4 levels of priority support
• 3 stage congestion management scheme
• No single point of failure
• No restrictions on slot usageT R
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
Port
1
Port
2
Port
8
T R
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
Port
1
Port
2
Port
8
T R
Switch
Processor
VLSI
Port
1
Port
2
Port
8
TT R R
Switch
Processor
VLSI
Switch
Processor
VLSI
High-speed
BUS
Interface
5. 5
Architectural OverviewArchitectural Overview
Completely distributed switch
architecture featuring:
• 4 levels of priority support
• 3 stage congestion management scheme
• No single point of failure
• No restrictions on slot usage
Cellenium Bus (2.56Gbps)
Passive 2.56 Gbps Cellenium bus provides
protocol-independent support for:
T R
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
Port
1
Port
2
Port
8
T R
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
Port
1
Port
2
Port
8
T R
Switch
Processor
VLSI
Port
1
Port
2
Port
8
TT R R
Switch
Processor
VLSI
Switch
Processor
VLSI
High-speed
BUS
Interface
• Ethernet
• Fast Ethernet
• Gigabit Ethernet
• FDDI
• ATM
6. 6
Architectural OverviewArchitectural Overview
Completely distributed switch
architecture featuring:
• 4 levels of priority support
• 3 stage congestion management scheme
• No single point of failure
• No restrictions on slot usage
Ethernet 4
Ethernet 3
Ethernet 2
Ethernet 1
MGMT / Power - 2
Token Ring /FDDI 2
Token Ring /FDDI 1
MGMT / Power - 1
Cellenium Bus (2.56Gbps)
Passive 2.56 Gbps Cellenium bus provides
protocol-independent support for:
T R
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
Port
1
Port
2
Port
8
T R
ASIC-Based
Switching
Processor
ASIC-Based
Switching
Processor
Port
1
Port
2
Port
8
T R
Switch
Processor
VLSI
Port
1
Port
2
Port
8
TT R R
Switch
Processor
VLSI
Switch
Processor
VLSI
High-speed
BUS
Interface
Plus Dedicated Buses:
• 4 Shared Ethernet buses
• 2 Token Ring/FDDI buses
• 2 Management/Power buses
• Ethernet
• Fast Ethernet
• Gigabit Ethernet
• FDDI
• ATM
7. 7
Benefit:Benefit:
Complete FlexibilityComplete Flexibility
Protocol Independent Cellenium bus supports multi-layer
switching between multiple technologies
- Optional 3LS module provides high-speed ASIC-based lP/IPX
layer-3 switching between any combination of:
Port Switched Ethernet
Group Switched Ethernet
Switched Ethernet
Switched Fast Ethernet
Gigabit Ethernet
FDDI
ATM
8. 8
Congestion Management:Congestion Management:
Conventional SwitchesConventional Switches
OTHER SWITCHES rely solely on buffers for passive
flow control
- When the buffers are flooded, packets are lost
If a packet is lost then frames have to be
re-transmitted
- Adds more traffic on the already heavily loaded network
- “Snowball” effect can bring the network to a halt
Switch “X”
Client 1
Client 2
PACKET LOSS
Server
10 Mbps
10 Mbps
10 Mbps
10 Mbps
9. 9
LANswitchLANswitch PlusPlus--
Architecture for Robust PerformanceArchitecture for Robust Performance
Bus Access Arbitrators make sure that
each port gets fair access to the bus2.56Gbps
Cellenium
Bus
T R T RT RT R
T R
Transmit
Buffer
Receive
Buffer
T R T R
PORT 1 PORT 2 PORT 3
PORT 4PORT 5PORT 6
3-Stage Active Congestion Management Scheme
- Stage 1: Normal Traffic
- Dedicated per port buffers provide “passive” flow control
- Smart arbitration scheme ensures “fairness”
10. 10
LANswitchLANswitch PlusPlus--
Architecture for PerformanceArchitecture for Performance
2.56Gbps
Cellenium
Bus
T R T RT RT R
T R
Transmit
Buffer
Receive
Buffer
T R T R
PORT 1 PORT 2 PORT 3
PORT 4PORT 5PORT 6
Stage 2: Heavy Traffic: Receive buffer filled
- Internal Active Congestion Control kicks in
I’m
full!
11. 11
LANswitchLANswitch PlusPlus--
Architecture for PerformanceArchitecture for Performance
2.56Gbps
Cellenium
Bus
T R T RT RT R
T R
Transmit
Buffer
Receive
Buffer
T R T R
PORT 1 PORT 2 PORT 3
PORT 4PORT 5PORT 6
Stage 2: Heavy Traffic: Receive buffer filled
- Internal Active Congestion Control kicks in
- Sends brief “back pressure” signal to sending port to temporarily
hold transmission.
- Does not impact ANY OTHER ports
- External end stations are not affected
- Normal operation resumes when receive buffer clears space
Internal
“Back Pressure”
Signal
12. 12
LANswitchLANswitch PlusPlus--
Architecture for PerformanceArchitecture for Performance
Stage 3: Extreme Loads!! Both Receive buffer
and Transmit buffers filled
- External Active Congestion Control kicks in
2.56Gbps
Cellenium
Bus
T R T RT RT R
T R
Transmit
Buffer
Receive
Buffer
T R T R
PORT 1 PORT 2 PORT 3
PORT 4PORT 5PORT 6
I’m
full!
I’m
also full!
13. 13
LANswitchLANswitch PlusPlus--
Architecture for PerformanceArchitecture for Performance
Stage 3: Extreme Loads!! Both Receive buffer and
Transmit buffers filled
- External Active Congestion Control kicks in
- Sends brief carrier sense signal to end-station to temporarily
hold transmission. Effectively extends buffer to end-station.
- Does not impact ANY OTHER ports
- Normal operation resumes when buffers clear space
2.56Gbps
Cellenium
Bus
T R T RT RT R
T R
Transmit
Buffer
Receive
Buffer
T R T R
PORT 1 PORT 2 PORT 3
PORT 4PORT 5PORT 6
14. 14
Benefit - Unparalleled PerformanceBenefit - Unparalleled Performance
Top Performer
“LANNET’s LANswitch turned in a
nearly perfect performance. Thanks
to backpressure and king-sized
buffers, it didn’t drop a single frame
in any of the overload tests. The
LANswitch also outran the
competition in the speed trials.”
15. 15
LANswitchLANswitch PlusPlus- Complete Multilayer- Complete Multilayer
Networking SolutionNetworking Solution
A family of scalable switching solutions
- Common architecture, modules, and features.
- Scalability from workgroup to backbone
- Total product integration=>total network solution
LET-36LET-10 LET-20
• 5-slot chassis
• 2.56 Gbps Cellenium Bus
• Redundant Power Supplies
• 10-slot chassis
• 2.56 Gbps Cellenium Bus
• Redundant Load Sharing
Power Supplies
• 18-slot chassis
• 2.56 Gbps Cellenium Bus
• Redundant Load Sharing
Power Supplies
Wiring Closet
Data Center
Backbone
Wiring Closet
Data CenterWorkgroup
16. 16
Benefit: Complete Multilayer/FaultBenefit: Complete Multilayer/Fault
Tolerant Network SolutionsTolerant Network Solutions
LANswitch Plus at the center as well as in the floor
Engineering
Finance
AccountingSales
Marketing
R & D
R & D
Operations
PCPC
LET-36LET-36
PCPC
PCPC
PCPC
PCPC
PCPC
PCPC
PCPC
PCPC
ServersServers
LET-20LET-20
LET-20LET-20
LET-10LET-10
LET-10LET-10
18. 18
LANswitchLANswitch PlusPlus ModulesModules
Workgroup Group Switch ModulesWorkgroup Group Switch Modules
1.28 Gbps2.56 Gbps
Cellenium
Bus
Port 1
Port 2
Port 3
Port 4
Port 5
Port 6
Port 7
Port 8
LANswitch™
LSE-108
2
4
6
8
1
3
5
7
RX
RF
C
TX
TF
IO
+A
SA
1
2
3
4
5
6
7
8
10Mbps
“Module”
LSE-108
Port 1
Port 3
Port 5
Port 7
Port 2
Port 4
Port 6
Port 8
1.28 Gbps2.56 Gbps
Bus
LANswitch™
LSE-208
2
4
6
8
1
3
5
7
RX
RF
C
TX
TF
IO
+A
SA
1
2
3
4
5
6
7
8
1 2
10Mbps
10Mbps
“Bank”
“Bank”
LSE-208
Cellenium
• one-slot module
• one 10Mbps switching engine
• bandwidth dynamically
allocated between 8 ports
• RJ-45
• full active congestion management
support
• full priority level support
• one-slot module
• two 10Mbps switching engines
• 8 ports, 4 per switch engine
• RJ-45
• full active congestion management
support
• full priority level support
• optional secured mode
• optional stand-alone mode
19. 19
LANswitchLANswitch PlusPlus ModulesModules
Workgroup Dedicated SwitchingWorkgroup Dedicated Switching
LSE-808
Port 1
Port 2
Port 3
Port 4
Port 5
Port 6
Port 7
Port 8
LANswitch™
LSE-808
2
4
6
8
1
3
5
7
RX
RF
C
TX
TF
IO
+A
SA
1
2
3
4
5
6
7
8
1
2
3
4
5
6
7
8
10Mbps
10Mbps
10Mbps
10Mbps
10Mbps
10Mbps
10Mbps
10Mbps
2.56 Gbps
Bus
Cellenium
• one-slot module
• eight 10Mbps switching engines
• 8 ports, each with dedicated switch engine
• RJ-45
• full active congestion management
support
• full priority level support
• optional stand-alone mode
20. 20
LANswitchLANswitch PlusPlus ModulesModules
Segment SwitchingSegment Switching
Port 1
Port 2
Port 3
Port 4
LANswitch™
LSE-808
2
4
6
8
1
3
5
7
RX
RF
C
TX
TF
IO
+A
SA
1
2
3
4
5
6
7
8
10Mbps
10Mbps
10Mbps
10Mbps
LSE-404S/FL
LSE-404S/FB
LSE-404S/RJ
Segment 1
Segment 2
Segment 3
Segment 4
LANswitch™
LSE-404S
2
4
1
35
RX
RF
C
TX
TF
IO
+A
SA
1
2
3
4
5
6
7
8
10Mbps
10Mbps
10Mbps
10Mbps
2.56 Gbps
Bus
Cellenium
2.56 Gbps
Bus
Cellenium
• one-slot module
• four 10Mbps switching engines
• 4 ports, each with dedicated switch engine
• ST, SMA, RJ-45
• full active congestion management
support
• full priority level support
• hardware backbone redundancy option
21. 21
LANswitchLANswitch PlusPlus ModulesModules
100BASE-TX Server Switching100BASE-TX Server Switching
LFE-4004
LANswitch™
RX
C
TX
A
STATUS
100Mbps
100Mbps
100Mbps
100Mbps1
2
3
4
LFE-4004
2.56 Gbps
Bus
Cellenium
• one-slot module
• four 100 Mbps switching engines
• 4 ports, each with dedicated switch engine
• RJ-45
• full active congestion management
support
• full priority level support
22. 22
LANswitchLANswitch PlusPlus ModulesModules
100BASE-TX/FX Segment Switching100BASE-TX/FX Segment Switching
LFE-4004
LANswitch™
RX
C
TX
A
STATUS
100Mbps
100Mbps
100Mbps
100Mbps1
2
3
4
LFE-4004S/RJ
LFE-4004S/F
2.56 Gbps
Bus
Cellenium
• one-slot module
• 512 MAC address support per port
• four 100 Mbps switching engines
• 4 ports, each with dedicated switch engine
• RJ-45 or SC connectors
• full active congestion management
support
• full priority level support
23. 23
LANswitchLANswitch PlusPlus ModulesModules
BackboneBackbone
155Mbps
LANswitch™
LSF-100
1
2
CONSOLE
OUT OF
BAND
OPTICAL
BYPASS
BT
TX
TF
PR
A/M
RX
RF
SR
SA
RESET
LANswitch
LSF-100
2.56 Gbps
Bus
Cellenium
• two-slot module
• 64K MAC address support
• DAS- 2 standard PMD MIC connectors
• SAS- 1 standard PMD MIC connector
• Inter-Switch Link (ISL) VLAN trunking support
• full active congestion management
support
• full priority level support
28. 28
Load-Balancing Using LEB-200Load-Balancing Using LEB-200
Data
Center
Headquarters Building 2
LEB-200
LSE-808
LFE-4004
LSE-808
LEB-200
LFE-100
LSE-808
LEB-200
LEB-200
LEB-200
LANswitchLANswitch
PlusPlus
LANswitchLANswitch
PlusPlus
LEB-200
LEB-200
• Scale bandwidth between switches in 200Mbps increments (full-duplex)
• Traffic is allocated to each LEB-200 on a per-slot basis
• Load balancing does not
require implementation
of VLANs
• Maintains full flow control
and priority level support
• Resilient: Traffic can be
re-assigned when a link
fails.
400Mbps = 2 x 200 Mbps
29. 29
LANswitchLANswitch PlusPlus --
Meets Current and Future RequirementsMeets Current and Future Requirements
10
BandwidthperUser
(Mbps)
Desktop Backbone
& High-speed
Segment
100
LSA Plus
LSE-808
LSE-404S
LSE-108
LSE-
208
LFE-4004
LFE-4004S
200
1000 LGE
3LS
LEB-200
LEB-200
Load balancing
multiple links
LSF-100
30. 30
LANswitchLANswitch PlusPlus Multilayer CapabilitiesMultilayer Capabilities
“...unlike some level-three switches
the LANswitch (Plus) does true IP and IPX routing
using standard IP routing protocols (RIP and OSPF)
and IPX routing using IPX-RIP...”
- Scott Bradner
31. 31
LANswitchLANswitch PlusPlus 3LS Module3LS Module
3LS
2.56 Gbps
2.56 Gbps
Bus
Cellenium
Provides IP/IPX switching in and between
VLANs
Performs true layer-3 switching
Industry 1st ASIC-based, multilayer IP/IPX switch module
Two-slot module plugs directly into LANswitch Plus Cellenium bus
Supports RIP, OSPF, SAP and ARP routing protocols
Bridges non-IP/IPX and non-routeable protocols
(DECnet, LAT, NetBios, Appletalk)
32. 32
LANswitchLANswitch PlusPlus and 3LS-and 3LS-
Multilayer Integration and FlexibilityMultilayer Integration and Flexibility
3LS
2.56 Gbps
2.56 Gbps
Bus
Cellenium
Integration into LANswitch Plus provides multilayer support for:
- VLANs
- Priority levels
- Active congestion control mechanism
Provides layer-3 switching over Ethernet, Fast Ethernet, Gigabit
Ethernet, FDDI, and ATM interfaces
Includes on-board SNMP agent for easy
configuration and management
Multiple 3LS modules can be meshed for
redundancy and load-balancing
33. 33
Inside The 3LSInside The 3LS
TransmitTransmit
BufferBuffer
High-Speed
CPU Route
Algorithm
Broadcast &
Unknown
Packets
Rx BufferRx Buffer
Low-PriorityLow-Priority
Shared
Bridging and
Routing
Tables
High-Speed
Silicon
Frame
Forwarding
Engine
2.56 Gbps Cellenium Bus
Rx BufferRx Buffer
High-PriorityHigh-Priority
Forwarding
Unit CPU
Unit
34. 34
Forwarding UnitForwarding Unit
ASIC-based to optimize performance for time critical tasks
Performs true layer-3 hardware routing of IP and IPX packets
- Examines the layer-3 information in EACH packet and bases its
path selection on this information.
- Modifies layer-3 information as necessary, such as decrementing
the TTL (Time-To-Live) field in each packet.
Bridges non-routeable packets
Performs Access Control per packet
- According to user- defined filters
Identifies and forwards special packets to Route Server Card
- ARP, RIP, OSPF and SAP packets
All ASICAll ASIC
BasedBased
35. 35
CPU UnitCPU Unit
RISC-based to optimize flexibility for non-time critical tasks
Handles execution of router protocols
- RIP, OSPF, SAP and ICMP
- Upgradeable to other protocols in future
- “One-hop” routing calculation
Calculates routes for new destinations
Handles ARP requests and updates the forwarding tables
Builds access control tables for filtering
Provides SNMP agent functionality
RISCRISC
BasedBased
36. 36
LANSwitchLANSwitch PlusPlus and 3LS:and 3LS:
Superb Multilayer PerformanceSuperb Multilayer Performance
Latency: 12us (micro-seconds)
- All forwarding is done in hardware
Forwarding rate: 1M pps IP or IPX
“LANNET’s LANswitch offers red-hot performance at Layers 2
and 3 thanks to its industry-first router implemented in silicon.” -
Data Communications, 5/97
“LANNET LANswitch 3LS routing performance results are as
good as they can theoretically be. It routes hundreds of
thousands packets per second of IP and IPX traffic at full wire
speed, with no packet losses.” - Scott Bradner
37. 37
Load SharingLoad Sharing
All 3LSs have IP/IPX interfaces for all nets
Sharing IPX load is done dynamically
- The loaded 3LS will be slower,naturally, in answering RIP
requests
IP traffic can be load-shared across multiple 3LS units
by:
- Defining different default routers for each IP subnet
- Configuring all other 3LSs, except of the preferred 3LS, to
send RIP packets with big metric
38. 38
Scalable layer-3 performance overScalable layer-3 performance over
multiple topologiesmultiple topologies
3LS
A single 3LS module supports multiple hubs over any
combination of Ethernet, FDDI, and ATM....
FDDI
ATM
Fast Ethernet
39. 39
Scalable layer-3 performance overScalable layer-3 performance over
multiple topologiesmultiple topologies
3LS
...or add an additional 3LS module to load-share traffic
and scale performance as requirements grow...
Gigabit
Ethernet
Load Sharing
Fast Ethernet
ATM
3LS
40. 40
Scalable layer-3 performance anywhereScalable layer-3 performance anywhere
in the networkin the network
3LS
...or distribute 3LS modules to provide multi-million pps
IP/IPX switching across the entire network
Gigabit
Ethernet
Load Sharing
Fast Ethernet
ATM
3LS
3LS3LS
41. 41
Priority & Video SupportPriority & Video Support
3LS supports two levels of priority:
- High and Low
3LS provides static support for IP multi-casts
- IP Multicast packets are forwarded to the VLANs as defined
by the user
- Planned future support for IGMP to enable dynamic multi-cast
capability.
42. 42
Security & Access ControlSecurity & Access Control
Access Control for the following attributes is done in
hardware :
- Station IP address
- IP subnet
- IPX network address
- Application/service type (Telnet, FTP, etc.)
The Access control policy can be either:
- forward , block, or block and report
No Performance DegradationNo Performance Degradation
- “(3LS) Throughput remained nearly identical whether IP
or IPX traffic was being filtered.” Tolly Group
43. 43
Layer-3 RedundancyLayer-3 Redundancy
Redundancy can be set between multiple 3LS
modules
- It is only necessary to configure the redundant 3LS for each
interface that is to be backed up.
- Redundant 3LS is configured with primary 3LS/router IP
address as backup address.
In case of router/3LS failure, the backup 3LS will
answer ARP requests destined to this backup
address
44. 44
Flexible Layer-3 Redundancy SupportFlexible Layer-3 Redundancy Support
3LS
Redundancy within a chassis, across multiple chassis,
or even to support external routers
- Active Back-Ups: Redundant 3LS units can still be used for active
load-sharing duties.
3LS
3LS3LS
3LS
3LS Unit A 3LS Unit B
3LS Unit C
“B” Backs-up “A”
Set Redundancy Across
Multiple Chassis
“C” Backs
up “B”
“A” Backs
up “C”
3LS
Unit E
3LS
Unit D
Set redundancy
within a chassis
3LS backs-up
an external router
External
Router
45. 45
Result:Result:
A Clearly Superior Multilayer SolutionA Clearly Superior Multilayer Solution
R
2.56 Gbps
Multilayer Integration
Full VLAN support
Four Priority Levels
Integrated Mgt Suite
Full SMON support
Load Sharing Capability
Hot Standby Capability
Multilayer Flexibility
250 Logical Interfaces Supporting
Ethernet, Fast Ethernet
Gigabit Ethernet,
FDDI, ATM
2.56 Gbps
Multilayer Compatibility
True Layer-3 Processing
Full RIP/ OSPF Support
Multilayer Performance
12 Micro-second Latency
1M pps IP/IPX switching
Low latency variation
Scalability / Resiliency
Multilayer Security
True Layer-3 processing
Hardware-based filtering
Access List Firewalls
No Impact on Performance!
3LS3LS
46. 46
LANswitchLANswitch Plus-Plus-
VLANs that Span the CampusVLANs that Span the Campus
256 VLANs per chassis and end-to-end across the
entire network
VLAN trunking over Fast Ethernet, FDDI, and ATM
Montage Plus AVM simplifies VLAN
configuration/management
- auto-configuration of VLANs by IP sub-net or IPX network
- reconfiguration of VLANs for moves, adds, and changes.
Global VLANs enable access to shared resources
without requirement for layer-3 forwarding
47. 47
LANswitchLANswitch PlusPlus/3LS/VLANs-/3LS/VLANs-
A Winning CombinationA Winning Combination
Global Virtual LAN
Virtual LAN 2
IP subnet 2 LANswitch
Plus
IPX A
IPX B
IP subnet 4
IP subnet 1
IP subnet 3
E-mail
server
Web
server
3LS Module forwards
IP/IPX between VLANs
at 1M pps.
Virtual LAN 1
Global VLAN server and printer
resources are accessible from
ALL VLANs. No awkward third
party NIC scheme required.
3LS
Montage Plus AVM
provides automatic
configuration of layer-3
IP/IPX VLANs
LANswitch Plus platform
ties everything together
with its unique Switch
Monitoring (SMON)
capability
48. 48
"Imagine how hard it would be to coach a baseball
team if instead of being able to view the whole field
with all your players in position, you could see only
one player at a time. When monitoring a switched
network, you run into a similar difficulty because,
unlike in shared networks, you can only see one port
at a time. If you're running a [LANswitch], however,
you're in luck.."
- PC Magazine -- April 23, 1996
LANswitch Plus SMONLANswitch Plus SMON
49. 49
Switch MonitoringSwitch Monitoring
Proactive FaultProactive Fault andand Performance ManagementPerformance Management
How can I
• Become more proactive?
• Drill-down to rapidly pin-point problems?
• Get an overall sense of activity?
• Identify performance bottle-necks?
• Monitor conversations crossing the switch?
• Analyze trends on a switch, port & user basis?
How can I
• Become more proactive?
• Drill-down to rapidly pin-point problems?
• Get an overall sense of activity?
• Identify performance bottle-necks?
• Monitor conversations crossing the switch?
• Analyze trends on a switch, port & user basis?
50. 50
Answer:Answer:
SMON (Switch-MONitoring)SMON (Switch-MONitoring)
RMON compliant
- Host Statistics
- Host Top N
- Host Matrix
Switch related additions
- Switch Statistics.
- Port Statistics
- Port Top N
Enterprise SMON further extends SMON to provide traffic
monitoring simultaneously across MULTIPLE switches!
SMON is a major RMON WG (IETF) agenda item for 1997
SMON
RMON
Enterprise SMON
51. 51
What is SMON?What is SMON?
A unique integrated hardware/software architecture:
- Passive Cellenium Bus
- NMA-RS plug-in probe residing on bus with view to all traffic
- SMONMaster console application
TEXT
100
BaseTE
T
H
CelleniumCellenium
2.56Gbps Backplane2.56Gbps BackplaneTEXT
VIDEO
A
T
M
E
T
H
E
T
H
F
D
D
I
T
R
VIDEO
VIDEO
TEXT
TEXT
DATABASE
SHARING
TEXT
100
BaseT
Agent
+
SMON
NMA-RS
Gig
Eth
52. 52
What is SMON?What is SMON?
A uniquely upgradeable architecture
- NMA-RS module is independent of the switch modules
- Architecture completely separates monitoring from
switching, providing upgrade path for future monitoring
capabilities such as RMON II functionality
“As new management capabilities need to be added, LANNET
can simply add more software to SMON. Other vendors using a
distributed approach may have to update both hardware and
software on every switching module to add new management
capabilities, costing enterprises both time and money.”
-Gartner Group
53. 53
What does SMON provide?What does SMON provide?
A global view of all switched traffic
Powerful “Top-down”
monitoring
No impact on switching
performance
Switch Monitoring
Conversation
Monitoring
Port Monitoring
VLAN and ELAN Monitoring
Enterprise Switch Monitoring
User Monitoring
“...since switching and monitoring aren't vying for the same resource, you can
constantly monitor the switch without degrading performance.”
- Gartner Group
54. 54
Enterprise Switch MonitoringEnterprise Switch Monitoring
Enterprise Switch StatisticsEnterprise Switch Statistics
View traffic patterns simultaneously across multiple
switches
Spot out-of-line conditions
anywhere in the network
Balance utilization
across the entire
network
55. 55
Switch MonitoringSwitch Monitoring
Switch StatisticsSwitch Statistics
Measure utilization across the entire switch
Measure overall QoS traffic allocation
Spot out-of-line conditions within the switch
Drill-down within the switch from here
56. 56
VLAN/ELAN MonitoringVLAN/ELAN Monitoring
Switch StatisticsSwitch Statistics
Compare traffic within all of your VLANs simultaneously
Identify bottlenecks and which VLANs are being under-
utilized
Network baselining capability enables you to optimize
VLAN performance
Make informed decisions
on whether to relieve
bottle-necks by adding
bandwidth or changing
traffic patterns
57. 57
Port MonitoringPort Monitoring
Port Statistics and Port TopNPort Statistics and Port TopN
Trace problems to individual ports on the switch
Determine traffic utilization per port/link
Spot bottle-necks and drill-down to analyze traffic
patterns within a port
Plan for future growth
58. 58
User MonitoringUser Monitoring
Host and Host TopNHost and Host TopN
Drill-down to identify hosts on particular ports
Identify heavy host users or problem end-stations
Look at traffic statistics for each host to identify
abnormal conditions
Study usage patterns
59. 59
Conversation MonitoringConversation Monitoring
Host MatrixHost Matrix
View information on active sessions between all host pairs on
the Cellenium bus
- Changes color based on relative traffic levels
- Display of high error rates
Drill down to monitor conversations
of heavy users or problem
end-stations
Measure Server usage by host
Analyze communications patterns
to optimize VLAN performance
60. 60
“SMONMaster gives you a bird's-eye view of your network,
providing such graphical statistics as which hosts on the
network are communicating and how much, which ports see
the most traffic and at what time of day, and whether any
ports are reporting error packets. This just isn't possible
with any other switch-monitoring software.”
“The ability to monitor and analyze network statistics from
all ports at once or to get an idea of traffic trends can be the
key to optimizing limited network resources and improving
performance.”
- PC Magazine
The Verdict?The Verdict?
““LANNET hits Switch-Monitoring Home Run”LANNET hits Switch-Monitoring Home Run”
61. 61
Benefit:Benefit:
“More Management....Less Money”“More Management....Less Money”
Enterprise SMON and unique drill down capabilities enable you
to optimize your network, not just configure it
Real-time traffic monitoring enables pro-active network
management, vs. after the fact.
Greater level of traffic reporting proivdes greater added value
without the need for expensive external analyzers
Superior fault tracing capabilities ensure network uptime
63. 63
LANswitchLANswitch PlusPlus--
Architecture for Resilient NetworksArchitecture for Resilient Networks
Resilient Links
- Hardware redundancy within a module providing recovery
time in micro-seconds
- Software redundancy within a hub providing
redundancy support across mixed technologies
recovery time of 10-20 seconds
- Software redundancy across multiple hubs providing
redundancy support across mixed technologies
recovery time less than 30 seconds
64. 64
Floor N
Data
Center
Floor 1
LANswitchLANswitch PlusPlus--
Hardware and Software Link RedundancyHardware and Software Link Redundancy
FDDI Primary
Link
Fast Ethernet
Redundant Link
Redundancy across
multiple topologies
(recovery 10-20 seconds)
Intra-module hardware
redundancy
(micro-seconds recovery)
Redundancy across
multiple hubs
(recovery <30 seconds)
Primary
Secondary
Primary
Secondary
65. 65
Benefit-Benefit-
Robust Network InfrastructureRobust Network Infrastructure
End-to-end resilience
No single point of failure
Redundancy across multiple topologies
leverages existing infrastructures
Rapid recovery times
66. 66
Conclusion- What Sets LANswitchConclusion- What Sets LANswitch PlusPlus
Apart from the CrowdApart from the Crowd
Resilience
• Distributed Architecture
• Hot-Swappable Modules
• Redundant Links
• Redundant Power Supplies
Multilayer Performance
• 1-2M pps IP/IPX Switching with
RIP/OSPF Support
• >2M pps Aggregate Throughput
• Traffic Prioritization (4 levels)
• Active Congestion Management
• Low Latency, Latency Variation
Flexibility and
Investment Protection
• Media Independent Architecture
• No slot restrictions
• Fast Ethernet Load Sharing
• Scalability to Gigabit Ethernet
• Multiple Topology Layer-3 Support
• VLAN Trunking across Multiple
Media
Superior Integration
• Multilayer VLAN Support
• Multilayer Priority Support
• Integrated Network Management
• Integrated Cell/Frame Monitoring
• Integrated ELAN/VLAN Mgt
• End-to-End VLANs
• End-to-End Congestion Control
SMON
• RMON Compliant
• Unique Switch Level Traffic
Monitoring
• Drill-Down Capability
• No Impact on Performance
Complete Family
• LET-10
• LET-20
• LET-36
• Visage Family
Editor's Notes
The RS card is responsible for all non-time critical tasks of the LRE. That includes handling of the routers protocols (RIP, OSPF, SAP and ICMP), routing calculation for new destinations, ARP handling and update of the forwarding tables, SNMP agent functionality, building of the required access control tables, interface with the box agent (via the SPI bus and the 68HC11 micro-controller), interface with terminal for set-up and basic control of the module and handling of access control (not handled by the forwarding card).
Splitting the traffic
As mentioned above, all the 3LSs will have IP/IPX interfaces to each net.
The IPX load is likely to be shared between the 3LSs naturally, as the loaded 3LS will probably be slower in answering RIP requests.
This is not so for IP. In order to help split the IP load between the different 3LSs, the user should divide the IP nets (or the VNs) between them. Each IP (sub)net will be assigned to one 3LS, whose address should be configured as the default router to the IP stations on that net that require configuration (e.g. PCs). The other 3LSs, that also have interfaces to that net, should be configured to send in RIP only the default route with big metric. This will lead stations that learn the router&apos;s address dynamically from RIP (e.g. SUNs) to use the preferred 3LS.
Similarly, if there are IP stations on that net that listen to ICMP router advertisement messages, all 3LSs except the preferred one should be configured to send these advertisements with lower preference level.
This way of dividing the nets between 3LSs, and configuring RIP to send big metric accordingly, is recommended in order to minimizes the number of cases where traffic between two ports of the same switch is routed through an 3LS on another switch, unnecessarily crossing the backbone twice.
IP nets that reside on VNs of one switch should be assigned to the 3LS located on that switch, and the other, remote 3LSs should be configured to only advertise themselves to these nets with big metric.
The dashed lines in the figure above represent IP interfaces, over which the router sends RIP with big metric. Thus, stations will send packets to the router on their switch and the packets will only cross the backbone if they really have to get to the other side.
This will not work for VNs that span multiple switches.
This will not work for IPX. In order to direct stations to use the preferred 3LS, the others have to not answer RIP requests. This will hurt the redundancy, though.
The 3LS support the emerging IP Multicast protocol for video conferenceing and forward these packets to the right VLANs as defined by the user. ( This capability isn’t supported by the current routers. The definition is static in the first ver.)
When the LNS receives an IP packet whose destination address is one of the configured class D addresses, the packet is forwarded to all ports of the corresponding configured list
Access Control
The LRE shall provide access control filters for the routed data. It shall have high speed filters, implemented in the hardware, that do not reduce the performance, and low speed filters, implemented in software. The low speed filters are full set of a regular router&apos;s filters and shall be used for the cases that are not handled by the hardware. It shall provide high speed filters for the following attributes:
- Station IP address
- IP subnet
- IPX network address
- Application/service type (Telnet, FTP, etc.)
Redundancy
In IPX, the LREs will naturally provide the same level of redundancy provided by any two IPX routers (e.g. OG), i.e. the level of redundancy provided by the IPX protocol itself. When an LRE fails the stations using it will automatically start using another LRE. The backup LRE must be configured to send RIP on the interface.
For IP, LRE will provide redundancy, such that if one of the LREs fails it is backed-up by another LRE, pre-configured to backup the first, without the user having to reconfigure all the IP stations on the net. The typical application will be two LREs configured to backup each other. Another application can be three LREs A, B and C, where A is configured backup to B, B to C and C to A.
An LRE that is configured to backup another will monitor the other LRE&apos;s status by listening to its RIP updates, and will thus be always in a mode of either inactive or active backup.
When active backup, the LRE will operate some special ARP behavior (see below).
The backup LRE has interfaces to all IP subnets, with its own IP addresses. Some IP stations learn their router&apos;s IP address dynamically from RIP or ICMP router advertisement. When the LRE they use fails, these stations will eventually (order of minutes) learn the IP address of the backup LRE and start using it.
The problem, unsolved usually, is with IP stations that are manually configured with the default router&apos;s IP address. These stations will never recover unless the user goes and re-configures the default router address in each of them. This burden is what usually bothers the users.
Such stations will continue sending ARP requests to the IP address of the failed router. This is also true for those smarter IP stations during the minutes till they learn the new router.
In order to satisfy that, the backup LRE will operate something similar to proxy-ARP. It will be configured with the first router&apos;s IP addresses as backup-address, and as long as it is active backup due to the failure of the first router it will answer ARP requests destined to those backup-addresses.
Although this provides automatic recovery, sessions will break, and the IP stations will lose connectivity till they send a new ARP request for the router, since till then they continue sending IP packets to the MAC address of the failed router. A typical station&apos;s ARP refresh time is in the order of 15 minutes, but there even are stations that never refresh their ARP cache until rebooted.
This will be solved by sending a &apos;loop-back ARP&apos;. When identifying the failure, the backup LRE will send an ARP request destined to itself, with the backup IP address in the sender&apos;s IP address field and its own MAC address in the sender&apos;s HW address field. Stations complying to the standard will update their ARP cache entry according to the source information in the packet, although the packet is not destined to themselves. Thus they will immediately direct their IP packets to the backup router&apos;s MAC address.
Network Managers, also need a top-down sense of ATM network activity so they can more easily anticipate or pin-point problems before resorting to more detailed analysis.
It is also important to have tools for performance analysis and trending to be able to balance utilization between switches or identify bottle-necks - like an overly busy port.
A recent study has shown that a major inhibitor to the wide-spread adoption of RMON is the fact that the majority of monitoring tools are considered too complex and initially too detailed. A top-down approach is far easier to work with.
These tools should be consistent with those used for Ethernet or Token Ring switch monitoring.
So how does Switch-Monitoring fit with the the RMON standard?
RMON (Remote Monitoring) is the internationally recognized standard for the detailed analysis of Ethernet and Token Ring media. Its roots are in the monitoring of shared-media. To accommodate the special needs of switch monitoring LANNET’s SMON is designed to embrace and extend RMON. The result is a powerful new architecture capable of collecting and displaying a top-down view of both cell and frame switched network activity.
SMON was recently proposed to the RMON working group in the IETF and accepted as a major agenda item for 1997. SMON is also complimentary with the emerging ATM Forum ATM-RMON standard (which at the time of writing is yet to be fully ratified). ATM-RMON is focused on the challenge of monitoring ATM switches, whereas SMON embraces the monitoring of all switches - ATM, Ethernet and Token Ring.
SMON is still the only solution today for true Ethernet switch monitoring.
Only the LANswitch Plus Cellenium bus architecture allows for complete top-down monitoring. Top-down switch management employs a multi-tier approach. If there is a problem on your switch, the first thing the network manager would look at is the general condition of the cellenium bus. From there you would want to look at the total traffic on each virtual lan and then further down to particular traffic between two users. SMON then provides further analysis down to the specific hub port and finally, to a particular user station attached to that port.
Using this top-down approach eliminates inefficient sequential searches characteristic of the other architectures. LANNET is the first to offer this top-down simultaneous view of ALL traffic on your LanSwitch network including ethernet, token ring, 100baseT, FDDI and ATM traffic.
Lets take a look at SMONMaster 2.0 in action.
This powerful view is NEW to SMONMaster 2.0 and extends the top down concept still further allowing Network managers to compare the relative utilization of different switches in a mesh and to rapidly respond to error conditions. (each IP address represents a Switch monitoring agent).
You can drill down from here to an individual switch and then to port, user or conversation views.
Note: The simultaneous view of multiple switches will not be implemented for ATM in this release - only for Ethernet (LANswitch).
This view is of the highest level bus and virtual LAN statistics.
The pie chart on the left displays the percentage of good packets versus error packets.
The pie chart in the middle displays the percent of unicast versus multicast and broadcast traffic. This is a critical statistic. Too much broadcast traffic affects all the equipment on your network. If SMON displays a large number of broadcast traffic, this may be a time to segment your network into Virtual LANs.
The third pie chart displays the percentage of traffic per priority.
The graph one the bottom half show various statistics that can be plotted to show traffic performance. One of the things that can be shown here is a traffic comparison of performance on each of your virtual networks. You can simultaneously look at all the virtual lans and instantly determine which vlans are under or over untilized.
Port and Port TopN can also be part of “top-down” problem resolution and can be used to show the busiest ports (in the case of ATM this usually also means the busiest links) which will help to balance traffic between ports/links.
Host statistics can be used to rapidly pinpoint the “top talkers” across the switch, making it easy to measure and highlight, heavy usage, or violations of allocated QoS. From here the Host Matrix view can be used to identify “who” the offending user is talking with.
The host matrix tools display information about conversations between all host pairs on the switch. The display is color coded for quick indication of active conversations, relative traffic levels, and diagnosis of error percentages.
By analyzing conversations between equipment, you can easily determine which users are using network equipment, such as routers or file servers, the most and allow you to partition or isolate that traffic using, for example, virtual LANs.
As another example, if you have many file servers on your network, the matrix would be able to instantly show you which servers are being over and which are bing under utilitzed.
In summary, the host matrix tool provides a unique view - allowing you to view traffic port to port . That is the way traffic actually flows through the switch.
Fully Distributed Switching Architecture - No single point of failure
Hot Swappable Modules
Redundant/ Load Sharing Power Supplies
Multiple Management Station Support with Distributed Management
On-board Configuration Stores
Separate Management, High-speed and Multiple Ethernet Buses