Date post: | 01-Apr-2015 |
Category: |
Documents |
Upload: | jalen-hudson |
View: | 213 times |
Download: | 0 times |
1© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
MAINS (Metro Architectures enablINg Subwavelengths)
Georgios Zervas (WPL, UEssex)
MAINS 1st EC Technical ReviewBrussels, March 24th 2011
WP4Experimental validation of the MAINS concept
2© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Contents
Brief WP4 Y2 summary
– Objectives– activities and – results
Technical insight on Y2 results
– TSON Node prototype– SLAE prototype (to be completed in Y3)– Virtual-PC over OPST Field Trial @ Primetel network
– Year 3 plans
3© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Main WP4 Year 2 objectives and plans for Year 3
[Y1] Identifying and purchasing the main required devices to support TSON.
[Y1] Map TSON architecture (T2.2) to TSON implementation rules.
[Y2] Design and implementation of TSON Node prototype ✓
[Y2] Field Trial of Virtual PC services over OPST on Primetel’s network. ✓
[Y2-Y3] Implementation of SLAE tool for sub-lambda (time) resource allocation. ✓
[Y3] TSON Network Test-bed. On-going
[Y3] TSON-OPST Transport/Data Plane Interworking. On-going
[Y3] TSON-OPST Interworking with extended GMPLS control plane. To start M28
4© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Activities breakdown
WP2
WP3
T 4.1 Implementation of Ring-
Mesh Interconnection Node and Metro Mesh
bypass Node
T 4.2 Field trial of OPST
ring demonstrating pc virtualisation
services
D.2.1T 4.3
End-to-end controlled Ring - Mesh integration
and performance evaluation
T 4.4Field trial of controlled
Ring-Mesh interconnected
network
D.4.1
D.4.4
D.4.3
D.4.2
D.3.2
WP1D.1.1
D.3.4
D.2.5
Y1/Y2 Y3Y3
Y2
5© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
WP4 Year2 Gantt Chart
Y2
M23Implementation of TSON Nodes
M23Field Trial of Virtual PC over OPST
M25Implementation of sub-lambda assignment element
M28OPST-TSON Interconnection
M33GMPLS controlled OPST-TSON Testbed
6© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Year 2 work summary: Objectives, activities and resultsObjectives and activities carried out in T4.1 (M12 –M25)
– Developing TSON Metro Node prototype for Edge and Bypass services• Layer 2 prototype system to deliver TSON services (Eth. parsing, allocation, aggregation, etc.) • Layer 1 prototype system to switch TSON data sets• Dynamic and programmable Layer 2 TSON reconfiguration
– Sub-Lambda Allocation Engine (SLAE) tool ending in Year 3 (M25)
Activities carried out in T4.2 (M21- M23):
– Field Trial demonstrating Virtual PC Service over OPST ring on Primetel’s network
Results– TSON Node (L2 and L1) prototype (T4.1, D4.1)– Field Trial of Virtual PC over OPST successfully performed (T4.2, D4.2)– SLAE prototype tool (T4.1, D4.5)
7© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Deliverables
DELIV. DESCRIPTION DEADLINE
D4.1 Implementation of OPST Metro Ring- OBST Metro Mesh interconnection Node and Mesh bypass Metro Node
M23(Year 2)
D4.2 Field trial of OPST ring demonstrating pc virtualization services
M23(Year 2)
D4.5 Implementation of sub-lambda assignment element
M25 (Year 3)
D4.3 Integration of controlled OBST Metro Mesh interconnected with OPST Metro Ring network
M28(Year 3)
D4.4 Field trial of GMPLS controlled Ring-Mesh interconnected network
M33(Year 3)
8© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
[D4.1 insight] Time-Shared Optical Network (TSON) Node Prototype
Software Platform to support :– Virtual PC application– Sub-wavelength enabled GMPLS Stack– Sub-lambda PCE for RWTA– Node interfaces
Layer 2 FPGA-based platform to support– 1x10GE and 2xTSON transceivers – Independent hardware sharing, and emulation.– Flexible time-slice aggregation, scheduling and optical
data formatting– Software-Hardware defIned Network (SHINE)
• Hitless reconfiguration from Packet-based (e.g. Ethernet) to TSON transport
Layer 1 optical nodes based on:– Only fast switches (PLZTs) for TSON support– Architecture on Demand (AoD) for flexible time and
frequency allocation.
9© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
MAINS TSON Node FPGA developments summary
HW module Features Lang. MAINS
Top Level≈5K code lines
Component instantiation, Glue logic VHDL/ (≈4200code lines)
MDIO Controller(Provides access to the configuration and status registers of PHY.)
VHDL (≈300 code lines)
I2C Controller(program I2C programmable XO/VCXO SI570 and synthesizer SI5368.)
VHDL (≈200 code lines)
Signal/pulse synchronization(Signal or pulse synchronization between different clock domains)
VHDL (≈300 code lines)
For TSON Metro node with one ingress/egress node, totally 30.2K code lines.
For TSON Metro node with two ingress/egress node, totally 48.6K code lines.
10© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
MAINS TSON Node FPGA developments summary(contd.)
HW module Features Lang. MAINS
Ingress node≈11.2K code lines
GTH transceivers (IP core)
VHDL/ngc (≈3200code lines)
10G Ethernet MAC (IP core)
VHDL/ngc (≈1200code lines)
Receiver Buffer (Receive Ethernet frames, keep good ones and drop bad ones )
VHDL (≈800 code lines)
Distribution block (Demux input Ethernet frames, put them to different FIFOs)
VHDL (≈1400 code lines)
Aggregation(Aggregate from Ethernet frame to TSON burst)
VHDL (≈2800 code lines)
Transmitter Buffer (Send burst out based on the Time-slice Allocation, Synchronization from RX clock to TX clock)
VHDL (≈1800 code lines)
11© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
MAINS TSON Node FPGA developments summary (contd.)
HW module Features Lang. MAINS
Eggress Node≈7.2K code lines
GTH transceivers (IP core)
VHDL/ngc (≈3200code lines)
10G Ethernet MAC (IP core)
VHDL/ngc (≈1200code lines)
Receiver Buffer(Receive TSON data sets )
VHDL (≈1500 code lines)
Seggregation (seggregate from TSON burst to Ethernet Frames)
VHDL (≈600 code lines)
Transmitter Buffer(Send Ethernet frame out when seggregation done, Synchronization from RX clock to TX clock)
VHDL (≈700 code lines)
12© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
MAINS TSON Node FPGA developments summary(contd.)
HW module Features Lang. MAINS
GMPLS communication≈6.4K code lines
GTH transceivers (IP core)
VHDL/ngc (≈3200code lines)
10G Ethernet MAC (IP core)
VHDL/ngc (≈1200code lines)
Receiver Buffer(Receive Ethernet frames, keep good ones and drop bad ones )
VHDL (≈800 code lines)
LUT/LUT_Update(Update the Time-slice allocation and PLZT switch Look-Up-Table )
VHDL (≈500 code lines)
Transmitter Buffer(Send Ethernet frame out to the server as a LUT update feedback, Synchronization from RX clock to TX clock)
VHDL (≈700 code lines)
PLZT Switch Control≈0.4K code lines
Control PLZT Switch based on the PLZT switch LUT VHDL (≈400 code lines)
13© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
[D4.5 insight] Sub-Lambda Allocation Engine Tool
Request from GMPLS CP: SLAE tool is being invoked by GMPLS:
– Network topology matrix – Number of wavelengths per link – Number of time slices per each
wavelength – Source node – Destination node – Bit-rate request – The number of path for KSP
Response back to GMPLS CP:
– The assigned path– Assignment matrix (wavelength &
time-slices)
14© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
MAINS SLAE developments summaryHW module Features Lang. MAINS
Sub Lambda Assignment Engine≈5K code lines
java main function java (≈1200code lines)
Wavelength and time slice allocation Algorithm java (≈600code lines)
Network information abstraction java (≈1400 code lines)
Routing data types and algorithms java (≈500 code lines)
Utilities functions java (≈900 code lines)
Outputs and statics java (≈400 code lines)
15© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
TSON Node/Network Testbed
Servers (2x10GE NICs each) for network control &Virtual PC Services
FPGA for L2 operation
PLZT(4x4s, 2x2s) switches
EDFAs
2 FPGA L2 platforms
FPGA SFP+ (80Km WDM) TX/RX
Back view: FPGA/(DE)MUX
16© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
[D4.2 insight] Field trial of OPST ring demonstrating pc virtualization services Field Trial and Setup
– Setup Planned– Topology– Node specifications– Configuration
Deployment and Operational Aspects– OPST deployment– Service Integration– Network Interfaces
Performance Evaluation and Quality of Experience– Use case scenarios– Scalability tests
17© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Field Trial: Virtual PC over OPST Ring on Field TrialPrimetel Setup Planned
19© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Field Trial: Virtual PC over OPST Ring on Field TrialActual Topology Setup
LINK TO VIDEO COULD GO HERE
20© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Field Trial: Virtual PC over OPST Ring on Field TrialIntune Configuration
Two virtualization servers are connected to Netgear switches, which in turn connect to Intune Beta nodes 1 and 3 in the symmetrical ring of 5 Km. span per segment.
The ring is formed by three Intune Beta nodes and one of them serves as a pass-through (node 2 in the picture).
The Netgear switches provide interface between the Server 1G ports and the 10G client interfaces on the InTune Beta nodes.
21© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Field Trial: Virtual PC over OPST Ring on Field Trial OPST Deploymentat Primetel Nicosia HQ
Figure 1: Two OPST nodes @ HQ
Figure 3: 10Gbit XFPs for Optical Switch to OPST nodes
Figure 2: Switches for Management over LAN
Figure 4: Research Servers 2 & 3
22© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Field Trial: Virtual PC over OPST Ring on Field TrialIntune Management System
A PC was also connected to the LAN to support configuration and management allowing for the following:
– When the three-node ring were powered up, the power levels had to be calibrated optically.– Bit error test, where PRBS test data is generated before action data traffic is allowed on the ring.– MAC addresses configuration.– Master Node Selection and Service Mode Initialization. – Virtual Connections Setup.
23© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Field Trial: Virtual PC over OPST Ring on Field TrialApplications Setup
A number of Virtual Machines were created on a single Laptop connected to the prototype network with real user interfaces.
With the Virtual Machine Application we were in position to create a number of Virtual Machines, shut them down, modify them and if necessary transfer them where necessary from one server to another.
Figure 1 (left) shows the login at MAINS Virtual PC Homepage Figure 2 (right) shows the user friendliness of the application where on is able to indicate on a map the
selected destination and hence server location. Server selection can be achieved through this interface. The orange smiley icon represents us, blue server is the selected server, and red servers are available servers
24© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Field Trial: Virtual PC over OPST Ring on Field TrialUse Case Scenarios
Quality of Experience of Scenario 1: In this use case scenario, the user is accessing his virtual machine from local server on the same network and hence experiences optimum performance.
Quality of Experience for Scenario 2: In this scenario the user is accessing his virtual machine from remote server located approximately 11km away. The quality level experienced by the user was not much different to the first scenario with very minor jitter experienced due to the transport layer protocol.
Quality of Experience for Scenario 3: In this scenario, we tested the transfer of a virtual machine from a remote to a local server, while a mobile user changed its point of access from one location to another. The handover time experienced by the user on the application level was mainly due to the copying time of the virtual machine caused by the transport layer protocol.
25© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Year 3 activities:TSON Network Testbed
TSON Network Testbed [ongoing work]– Connect multiple TSON Edge and Bypass Nodes together ()
OPST-TSON Interworking [ongoing work]– OPST nodes to be hosted at Uessex and trial end-to-end OPST-TSON
network (
Vertical Integration of OPST and TSON with GMPLS [ongoing work]– Software integration among GMPLS-PCE-XML– Control-Transport Plane integration
Workshop/Demo at ECOC 2013– Control Plane workshop organized by Juan Pedro Fernandez-Palacios– Remote demos
• Control plane
26© 2010 MAINS Consortium MAINS 2nd EC Technical Review, Brussels, March 29th 2012
Thank you