evo controller 8200/rnc - infinity supply
TRANSCRIPT
BY : DINESH MALHOTRA
EVO Controller 8200/RNC
Introduction- Evo Controller 8200/Multi
• The Evo Controller 8200/Multi introduces a common building practice, processing device and switch for both GSM BSC and WCDMA RNC. It contains one subrack of a WRAN Evo Controller 8200/RNC and two subracks of a GSM Evo Controller 8200/BSC in one cabinet.
• From a radio and transport network point of view, the RNC and BSC functions are completely separated and treated as separate logical nodes. However, the GSM BSC and WCDMA RNC applications are run on the same type of Hardware (HW). Thanks to the same type of HW all in one cabinet, fewer spare parts are needed and the floor space can be saved. The Evo Controller 8200/Multi cabinet can also be used for free expansion later, becoming either a pure RNC or a pure BSC cabinet.
• The common HW is as follows:• Cabinet (BYB 501) • EGEM2 subrack • Evo Processor Board (EPB) • Ethernet switches (SCXB and CMXB) • Power and Fan Module (PFM)• The EPB is the key enabler of the multi controllers. It can be moved from a BSC subrack to an
RNC subrack, or the reverse, without the need to be reprogrammed in any way - simple plug&play is sufficient.
Capacity and Benefit of Evo Controller• Evolved Controller ~ RNC or BSC• Higher Capacity (More signaling and more data traffic, approximately 5 times to RNC3820 R1)• Low Cost to the performance• All IP (Evo-ET for ATM would be introduced later)• W11B~Software release supported for EVO.• RNC and BSC may stay together in one Cabinet
• 20 Gbps Iub throughput (5 times to RNC3820 R1.1)---HW Ready for 50 Gbps• 5 times signaling capacity to RNC3820 R1.1—Scalable upto 10 times with SW change.
• HSPA Evolution with 100 Mbps peak-rate for downlink and upto 12 Mbps for Uplink.• 2304 cell 738 Iub (16592 cells in theory)• 155448 active users (in theory)• EVO-BSC capacity 4095 TRX---HW ready to support upto 8000 TRX.
• 65% production cost reduction to Gbps, compared with RNC3820 R1.1• 73% power reduction to Gbps, compared with RNC3820 R1.1 • Performance Processor of EPB = 1.27 * SPB3
• The EvoC 8200/RNC is built with one Main Subrack (MS) and up to two Extension Subracks (ESs), which are all housed in one cabinet. Apart from the incorporated Power and Fan Module (PFM), each subrack contains a cable shelf at the bottom. The connection field in the EvoC 8200/RNC is called Active Patch Panel (APP). A standard cabinet is equipped with two APPs located one above the other at the very bottom of the cabinet.
• Figure shows the layout of the EvoC
8200/RNC cabinet.
CABINET- BYB 501 with BFD 538
RNC ModulesRNC modules divide the EvoC 8200/RNC into smaller resource units. An RNC module consists of a processing unit on a Module Controller (MC) and a number of associated devices within an Evo Processor Board (EPB). The number of RNC modules can differ in the Main Subrack (MS) and in Extension Subracks (ES) as the node configuration is customized.
Subrack EquipmentThe MS and ES can each house up to 28 boards. The subracks can contain EPBs, System Control Switch Boards (SCXBs) and Common Main Switch Boards (CMXBs), as well as some Dummy Boards (DBs. The EGEM2 subrack has 28-slots where each slot is 15 mm wide. Each slothas a duplicated 1 Gbps- and 10/40 Gbps Ethernet connection. The total backplane switching capacity per subrack is 960 Gbps. Up to 3 EGEM2 subracks can be used, which gives a total of 84 slots for plug in units. Each subrack is self-contained and includes temperature controlled fans.
Main Subrack Software Configuration
20 Configurable slots for EPB_BLADE or EVO-ET
Extension Subrack Software Configuration
24 Configurable slots for EPB_BLADE or EVO-ET
EISL
CISL
Internal Node Connection
EVO-C Hardware Building Blocks
Future proof investment with few HW types
Optional
Concept of RNC Module in EVO Controller
RNC Module
GPB EPB
EVO Controller
SPB
SPB
SPBGPB
SPB
Plug in unitsThe Evo Controller 8200 has been designed to reduce the different types of plug in units in order to reduce operator cost for spare part handling and simplify maintenance. Each board is 225x265x15 mm.The plug in units can be replaced without ISP impact. All plug in units are either working in a 1+1 redundant mode or with n+1 redundancy. The Evo Controller is designed for telecom grade performance with an availability of 99.999% or better
Plug in units Evo Controller 8200/BSC
Evo Controller 8200/RNC
Evo Controller 8200/Multi
EPB1 x x x
CMXB3 optional x x
SCXB3 x x x
EvoET optional optional optional
APZ 212 60C x - x
APG43/2 x - x
NWI-E x - x
[
EPB1 • Evo Processor Board (EPB1) ROJ 208 394/1
– Generic processor board used for all RNC processing tasks. Each EPB1 board is equipped with two multi-core processors.
• The EPB combines the roles of GPB, SPB and ET-IPG (which are used in RNC 3820) on the same board.
– There are three different Software Allocations (SWAs) for EPB:
• EPB_C1 RNC Central Processor 1 placed only in Main Subrack in slots 3 and 25.
• EPB_C2 RNC Central Processor 2 placed only in Main Subrack in slots 4 and 26.
• EPB_BLADE_A Has three Module Controllers, two DC devices and one CC device on the primary processor, and one PDR and six DC devices in the secondary processor.
– The four (‘c1’ and ‘c2’) EPB boards located in the Main Subrack are used only for central main processing tasks (both 1+1 redundant).
– The rest of the EPB boards (up to 20 in the main subrack and up to 24 in the extension subracks) are used for the Blade (traffic processing) role.
EPB1 (cont..)• The EPB1 board has 2 processors with 8 cores = total 16 cores.
• In Evo Controller 8200, the roles on each core are pre-defined making it easy to predict capacity expansions. The capacity scales with each EPB blade installed. The pre-defined core configuration is:
• 3 cores for MC• 8 cores for DC• 1 core for CC• 1 core for PDR• 1 core for CPP programs• 2 cores for IP termination
– Every EPB1 blade is configured with MC, CC, DC and PDR roles all on the same board. Thus one individual call is handled within the same EPB board, reducing signaling between boards in the Evo controller.
RNC Software deploymentC1 (1+1)-Equal to 3820 C1 + 2 x SCTP FE-Load balanced between cores (except JVM and Bare Metal)
“Blade” (3 – 68)
C2 (1+1)- 2 x SCTP Front End- Central device handling & UE reg- RFN server (moved from TUB)
Primary processorBDH + CPP
IP (bare metal)
CC device
RNC Module + RANAP
RNC Module + RNSAP
RNC Module + PCAP
DC device
DC device
Secondary processorPDR device + CPP
IP (bare metal)
DC device
DC device
DC device
DC device
DC device
DC device
SCXB3• System Control Switch Boards (SCXB3) ROJ 208 395/1
– SCXB is used to carry node internal control signaling and manages system clock distribution, and connections between EGEM2 subracks for control traffic.
– There are two SCXBs in every subrack belonging to two physically separated LANs.
• Each subrack contains a redundant SCXB pair, in slots 1 and 27 • All device boards and switch boards in a subrack are connected to both
SCXBs through a 1Gbps backplane connection. .
– The MS SCXBs are connected to the ES SCXBs by 10Gbps front panel Control Inter Switch Links (CISLs) - (slot 1 to slot 1 and slot 27 to slot 27).
– Functionally comparable to SCB and TUB in RNC 3820
SCXB
CMXB
CMXB
SCXB
SCXB
CMXB
CMXB
SCXB
CMXB
SCXB
CMXB
SCXB
EPBEPB
EPB1G1G
10G 10G
1G1G
1G1G
• SCXB3 Connectivity
Ethernet Cabling – SCXB3
MS – Cabling recommendations: A: to ES1 (A) (10G) B: to ES2 (A) (10G) C-G: Spare (10G) H-I : Not Used (1G)
J: 2048kHz or 10mHz sync ref
K: Not Used
L: Management - RS232
M: DBG – ETH
N: 2048kHz or 10mHz sync ref (QMA con)
ES1 and ES2 – Cabling recommendations: A: to MS (A & B) (10G) B: Spare (10G) C to H: Not used
CMXB3 • Common Main Switching Board 3 (CMXB3) ROJ 208 392/1
– Carries user plane data and node external traffic.
– Each subrack contains a redundant CMXB pair, in slots 2 and 28. All CMXB boards will be interconnected in a double star topology with Ethernet Subrack Links (ESL) – slot 2 to slot 28, slot 2 to slot 2 and slot 28 to slot 28.
– The board has 24 (10Gbps) Backplane ports which reach all (non switch) device boards within the same subrack.
– Also has 8 front panel (1 or 10Gbps) ports that can be used to connect to other CMXBs inside the node or as external ports for outgoing traffic. Ports A –D are HW prepared for 40Gbps.
– Functionally comparable to CMXB in RNC 3820
EvoC 8200/RNC
MS – Cabling recommendations: A: to APP (Iub) (10G, HW prepared for 40G) B: to ES1 (10G, HW prepared for 40G) C: to ES2 (10G, HW prepared for 40G) D: Cross-link in MS (10G, HW prepared for 40G) E: to APP (IuPS/CS, Iur) (10G)
F: to APP (10G)
G: to APP (10G)
H: to APP (10G)
ES1 and ES2 – Cabling recommendations: A: to MS (10G, HW prepared for 40G) B: cross-link in ES (10G, HW prepared for 40G) D to H: Not used
CMXB
CMXB
CMXB
CMXB
CMXB
CMXB
EPBEPB
EPB
4x10G4x10G
10G
10G
10G
10G
10G
10G
10G
APP APP4x10G4x10G
10G10G
Ethernet Cabling – CMXB3
EvoET• The EvoET board is an optional board to handle ATM
transport for Iub, Iur, and Iu-CS in the RNC and Packet Abis over TDM transport in the BSC.
• The EvoET logically terminates the node external transmission interfaces. Different versions exist for different standards and transmission speeds e.g.
STM-1/VC4 and STM-1/VC12. The board uses a 1+1 redundancy principle.
• There are 8 ports per EvoET. The number of EvoETs required and number of activated ports per board is dependant on the transport requirements.
• The EvoET uses the same board positions as the EPBs. The maximum no decapacity is therefore reduced when ATM- or TDM transport is required
CMXB3 allow 24 * 10GE backplane ports3 * 10GE + 5 * 40GE + 4 * 1GE front ports
SCXB
10GE
10GE
10GE
10GE
1GE
1GE
SyncExtA
DbgAPP
RPBS
SCXB
10GE
10GE
10GE
10GE
1GE
1GE
SyncExtA
DbgAPP
RPBS
CMXB3
40GE
40GE
40GE
40GE
10GE
10GE
10GE
40GE
Dbg
CMXB3
40GE
40GE
40GE
40GE
10GE
10GE
10GE
40GE
Dbg
SCXB provides 28x1G backplane ports and 4x10G + 2x1G front ports
White fronts and new LEDs
1GE
1GE
1GE
1GE1GE
1GE
1GE
1GE
Evo-C Provides 10G and 1G Ethernet to all Slot Positions
APP • Active Patch Panel (APP) KDU 137 557/4
– The APP converts the electrical 10GE interface of CMXB3 to various types of 1 GE and 10 GE optical interfaces by use of different types of SFPs. A 1000BaseT SFP electrical interface is also provided.
– The Active Patch Panel (APP) is the main physical point of connection for transmission. There are two APP units placed at the bottom, below the Main Subrack (MS).
– The APP provides O&M connections for a service terminal where one Ethernet connection and one RS232 connection can be used to connect to the Evo Controller.
APP – EXTERNAL IP INTERFACES• EVO external interfaces are very similar to what we have for RNC3820 with regards to
IP/cabling for OAM and bearers.– The only differences are related to the slots used for the internal connection points in the
subracks.
There are 3 pairs of External IP connection points:
•1. One pair of OAM Ethernet connections2. One pair of 1 or 10 GigE connections to carry
• Iub over IP traffic (User Plane and Control Plane) • 3. One pair of 1 or 10 GigE connection to carry
• IuCS, IuPS and Iur over IP traffic (bearer and signaling)
IuCS, IuPS & Iur
SFP - 32
SFP - 42
RJ45 - 61
BYB 501 with BFD 538 and CAS/CAXB
CAS Subrack with one CAXB board disassembled
APP is used in first release.Cabinet switch will be introduced in later release.
Cabinet Switch
SW Description - overview
›W11B EVO (W11.2) Software Changes:– The HW platform is completely new. In between the feature SW, (which is the same
as W11B), and the HW there is a middleware layer (e.g. in device & resource handling and user plane handling) that adapts the HW to the feature SW layer. That layer is updated in order to adapt the SW to the blade concept and the IP infrastructure.
– Platform: CPP9 (vs CPP8 for 3820)
›Features:– Same features as W11B– Dynamic Iu/Iur Signaling – This is a new feature required for EVOc
›Maintenance/Merging:– W11.1 (W11B) & W11.2 (W11B EVO) Merging in W12B– Standard TR mapping process between releases
Hardware ComparisonRNC 3820 EVO-C 8200
Subrack High Capacity Subrack EGEM2 Subrack
Subrack switching 200 Gbps 960 Gbps
Capacity 8 Gbps Iub throughput Hardware prepared for 20 Gbps
20 Gbps max. Iub throughput HW-prepared for 50 Gbps
Capacity (2) Max. 32 module MPs and 40 SPBs In a max. configuration of 68 EPB blades: 204 MCs, 68 PDR devices, 68 CC devices and 544 DC devices
Backplane Switching 10 Gbps HW ready for 40 Gbps
IP connectivity ETIPG, CMXB EPB, CMXB
Processing GPB, SPB EPB
Local sync source and timing unit
TUB Timing unit on the SCXB
Optical Indicators Three: red (fault), yellow (information) and green (operation)
Four: a new blue optical indicator (maintenance)
Router Path Supervision (RPS)
Supported Not supported. External mechanisms, such as Virtual Router Redundancy Protocol (VRRP), must be used for router supervision.
Software Comparison
RNC 3820 EVO-C 8200IP configuration ET board required to
terminate external IP trafficBoth node external and node internal IP traffic is terminated directly on the EPB
Dimensioning Control Plane (CP) vs. User Plane (UP) dimensioning, CP increased with GPBs, UP increased with SPBs.
Both CP and UP increased in a linear fashion with the number of EPB blades installed.
IP Termination Comparison
RNC 3820 EVO-C 8200
Iu and IurControlPlane
Mono signaling stack configured as arobust Sctp, where the RPU is configuredwith both active and passive slot and with amulti-homed IP host, where the physicallinks are on two separate ET-IPG boards.
Horizontally distributed signaling stackconfigured as a non-robust Sctp. One SCTPFront End (FE) executes on each C1 and C2EPB (four FEs in total).
IubControlPlane
Configured on ET-IPG. Redundancy isprovided by using IpAccessHostEt andsharing one IP host with Iub UP.
Configured on one EPB only; no ET boardredundancy.
Iu-PSUserPlane
Configured with one IpAccessHostSpbinstance for each PDR device.
Configured with IpAccesssHostPool andIpAccessHostEt.
Iu-CS andIur UserPlane
The same IpInterface, IpAccessHostEt,IpAccessHostPool instances can be usedboth for the Iu-CS and Iur unless differentVlans/subnetworks are required to separatethe planes.
Configuration with IpAccesshostPool issimilar, but the underlying implementation ischanged. The pool is used to identify whichinterface a certain IpAccessHostEt belongsto, but the selection of IP host is based onwhich EPB is handling the call.
Processing ComparisonProcessing RNC 3820 EVO-C 8200 GPB SPB EPB Blade
RANAP/RNSAP/SCCP
C2 Board
Distributed to the blades
Control plane, Iub links
One dual-coreprocessor withone RNCmodule oneach core
Three Module Controllers (MCs) on the primary processor (PIU)
User plane, devices
Three dual-coreprocessors andPDR, CC, DCdevices arrangedinto four differentSPB_TYPES
Two DC devices and one CC device on thePIU, Six DC devices and one PDR deviceon the secondary processor (PIU Device)