UNIVERSITY OF CALIFORNIA
Santa Barbara
Hybrid Adaptation Layers for High Performance Optical Packet Switched IP
Networks
A Dissertation submitted in partial satisfaction of the
requirements for the degree Doctor of Philosophy
in Electrical and Computer Engineering
by
Suresh Rangarajan
Committee in charge:
Professor Daniel J. Blumenthal, Committee Chair
Professor John E. Bowers
Professor Kevin Almeroth
Professor Upamanyu Madhow
March 2006
UMI Number: 3206436
32064362006
Copyright 2006 byRangarajan, Suresh
UMI MicroformCopyright
All rights reserved. This microform edition is protected against unauthorized copying under Title 17, United States Code.
ProQuest Information and Learning Company 300 North Zeeb Road
P.O. Box 1346 Ann Arbor, MI 48106-1346
All rights reserved.
by ProQuest Information and Learning Company.
The dissertation of Suresh Rangarajan is approved.
_________________________________________
Professor Upamanyu Madhow
_________________________________________
Professor Kevin Almeroth
_________________________________________
Professor John E. Bowers
_________________________________________
Professor Daniel J. Blumenthal, Committee Chair
March, 2006
Hybrid Adaptation Layers for High Performance Optical Packet Switched IP
Networks
Copyright © March, 2006
by
Suresh Rangarajan
iii
ACKNOWLEDGEMENTS
This thesis would not have been possible without the guidance, help and
encouragement of many people.
My advisor, Prof. Dan Blumenthal has provided me with the freedom to research by
granting me access to the Optical Communication and Photonic Networks lab. I
would also like to thank him for sharing his vision of optical networking with me and
allowing me to be a part of what I see as some of the most interesting engineering
research in this field.
I would like to thank my defense committee members both for taking time to read
through drafts of my thesis and for providing me with useful recommendations; Prof.
Kevin Almeroth for going through my work from a network viewpoint, Prof.
Madhow for his insightful comments on the high level perspective of my research and
Prof. Bowers for suggesting historical research in similar areas and for reviewing my
work from an optical communications view.
I would like to thank funding organizations Cisco and DARPA-Dimi for financing the
research presented in this thesis. Without their continued support for this work, this
thesis would have lost a significant part of its focus.
I would like to especially thank members of Cisco - Paul Donner, Russell Gyurek and
Farzam Toudeh-Fallah for their invaluable suggestions during most part of this
research.
iv
I would like to thank my colleagues at the Optical Communication and Photonic
Networks and Prof. Bower’s group for their collaboration, encouragement and advice.
I would especially like to thank Henrik Poulsen and Lavanya Rau for guiding me
through this research and for making it a pleasure to work in the lab. Roopesh Doshi,
Ramesh Rajaduray, Vikrant Lal, Marcelo Davanco, Emily Burmeister, Joe Summers,
David Wolfson, Zhaoyang Hu and others have all made my experience in this group
unforgettable both academically and socially.
I would like to thank my friends Jorn Helms, Mojtaba Ghodsi, Giulia Casavecchia,
Frank Alvarado, Graham Rhodes, Mahesh Khale and many more who have made
living in the beautiful city of Santa Barbara memorable.
Finally, I would like to thank my mom Vijayalakshmy Rangarajan, my dad
Rangarajan Gopala, my sister Priya Rangarajan and my brother-in-law Narayanan
Seshadri for their constant support, excellent and timely advice and helping me in
every way to complete my thesis.
v
VITA OF SURESH RANGARAJAN March 2006
EDUCATION
Bachelor of Engineering in Electrical and Communication Engineering from the
University of Madras, Chennai, India in April, 1999
Masters of Science in Electrical and Computer Engineering from the University of
California, Santa Barbara in June 2001
PUBLICATIONS
[1] Scalable All-Optical Compression/Decompression of Variable Length (40 to
1500 byte) Packets, Suresh Rangarajan, Henrik N. Poulsen and Daniel J.
Blumenthal – Optical Fiber Conference OFC’2006(accepted for presentation and
publication)
[2] All-Optical Packet Compression of Variable Length Packets from 40 to 1500
bytes using a Gated Fiber Loop, Suresh Rangarajan, Henrik N. Poulsen, Daniel J.
Blumenthal, IEEE Photonic Technology Letters, Jan 2006.
[3] Burst Mode 10Gbps Optical Header Recovery and Lookup Processing for
Asynchronous Variable-Length 40Gbps Optical Packet Switching, Henrik N.
Poulsen, David Wolfson, Suresh Rangarajan, Daniel J. Blumenthal, Optical Fiber
Conference OFC’2006(accepted for presentation and publication)
[4] Scalable All-Optical Packet Decompression of Variable Length Packets from
40 to 1500 bytes using a Gated Fiber Loop, Suresh Rangarajan, Henrik N. Poulsen,
Daniel J. Blumenthal, IEEE Photonic Technology Letters, (to be submitted)
vi
[5] Synchronizing Optical Data and Electrical Control Planes in Asynchronous
Optical Packet Switches, David Wolfson1, Henrik N. Poulsen1, Suresh Rangarajan1,
Zhaoyang Hu1, Daniel J. Blumenthal1, Garry Epps2, David Civello2; 1Univ. of
California at Santa Barbara, USA, 2Cisco Systems, USA, Optical Fiber Conference
OFC’2006(accepted for presentation and publication)
[6] End-to-End Layer-3 (IP) Packet Throughput and Latency Performance
Measurements in an All-Optical Label Switched Network with Dynamic For-
warding, Suresh Rangarajan1 , Henrik N. Poulsen1 , Paul G. Donner2, Russell
Gyurek2, Vikrant Lal1, Milan L. Masanovic1, Daniel L. Blumenthal1; 1:Univ. of
California, Santa Barbara, USA, 2:Cisco Systems Inc., USA, OWC5, Optical Fiber
Conference OFC’2005
[7] Performance of a Label Erase and Wave-length Switching Sub-System for
Layer-3 All-Optical Label Switching Using a Two Stage InP Wavelength
Converter, Henrik N. Poulsen, Suresh Rangarajan, Milan L. Masanovic, Vikrant Lal,
Daniel J. Blumenthal; Univ. of California at Santa Barbara, USA, OTuC2, Optical
Fiber Conference OFC’2005
[8] All-optical contention resolution with wavelength conversion for
asynchronous variable-length 40 Gb/s optical packets, Rangarajan, S.; Zhaoyang
Hu; Rau, L.; Blumenthal, D.J., Photonics Technology Letters, IEEE,
Volume:16, Issue:2, Feb.2004, Pages:689 – 691
[9] Analog performance of an ultrafast sampled-time all-optical fiber XPM
wavelength converter, Rau, L.; Doshi, R.; Rangarajan, S.; Yijen Chiu; Blumenthal,
vii
D.J.; Bowers, J.E.; Photonics Technology Letters, IEEE ,Volume: 15 , Issue: 4
, April 2003 Pages:560 – 562
[10] An all-optical bufferless multiwavelength sorter for 40 Gb/s asynchronous
variable-length optical packets, Hu, Z.; Rangarajan, S.; Rau, L.; Blumenthal, D.;
Optical Fiber Communications Conference, OFC’2003 , 23-28 March 2003
Pages:131 - 133 vol.1
[11] Optical signal processing for optical packet switching networks, Blumenthal,
D.J.; Bowers, J.E.; Rau, L.; Hsu-Feng Chou; Rangarajan, S.; Wei Wang; Poulsen,
K.N.; Communications Magazine, IEEE, Volume: 41, Issue: 2 , Feb. 2003 Pages:S23
- S29
[12] Low power penalty 80 to 10 Gbit/s OTDM demultiplexer using standing-
wave enhanced electroabsorption modulator with reduced driving voltage, Hsu-
Feng Chou; Yi-Jen Chiu; Rau, L.; Wei Wang; Rangarajan, S.; Bowers, J.E.;
Blumenthal, D.J.;Electronics Letters ,Volume: 39 , Issue: 1 , 9 Jan 2003 Pages:94 –
95
[13] Two-hop all-optical label swapping with variable length 80 Gb/s packets
and 10 Gb/s labels using nonlinear fiber wavelength converters,
unicast/multicast output and a single EAM for 80- to 10 Gb/s packet
demultiplexing, Rau, L.; Rangarajan, S.; Blumenthal, D.J.; Chou, H.-F.; Chiu, Y.-J.;
Bowers, J.E; Optical Fiber Communication Conference, OFC’2002 , 17-22 March
2002 Pages:FD2-1 - FD2-3
viii
[14] All-optical add-drop of an OTDM channel using an ultra-fast fiber based
wavelength converter, Rau, L.; Rangarajan, S.; Wei Wang; Blumenthal, D.J.;
Optical Fiber Communication Conference, OFC’2002 , 17-22 March 2002 Pages:259
– 261
[15] Optical packet switching and associated optical signal processing,
Blumenthal, D.J.; Bowers, J.E.; Chiu, Y.-J.; Chou, H.-F.; Olsson, B.-E.; Rangarajan,
S.; Rau, L.; Wang, W.; 2002 IEEE/LEOS Summer Topicals , 15-17 July 2002
Pages:TuG2-17 - TuG2-18
[16] Design of Impedance matching Microstrip line for Opto-microwave
applications, S.Rangarajan, A. Selvarajan et.al. National Communications
ix
ABSTRACT
Hybrid Adaptation Layers for High Performance Optical Packet Switched IP Networks
by
Suresh Rangarajan
With the increasing demand for network bandwidth, today's electronic routers
struggle to keep up and are fast reaching their physical limitations. All-Optical packet
switching offers an alternative approach to forwarding traffic by keeping the payload
in the optical domain throughout the network while forwarding is done based solely
on the optical header. Advantages of such an approach include a scalable increase in
payload bit rate, packet forwarding rate and reduction in power consumption and
physical size.
Current research in optics has yielded several integrated and non-integrated
technologies leading towards all-optical packet forwarding. Research on an
adaptation mechanism that achieves complete payload and frame decoupling and
thereby enabling direct IP over optics packet forwarding has however been limited.
Additionally, core packet forwarding involves statistical multiplexing of packet
traffic. Contention is a major issue during statistical multiplexing and has been
conventionally solved using Random Access Memory (RAM) buffers in electronics.
In the absence of Optical RAMs, a new technique of handling contention using fixed
length delay lines as buffering is required. This thesis focuses on research of a new
optical framing mechanism that enables all-optical statistical demultiplexing of IP
x
over optics at the core and optical processing mechanisms such as packet
compression and decompression to enable all-optical packet forwarding by statistical
multiplexing at the network core. Research on efficient adaptation of slow speed IP
traffic at the edge to enable and improve forwarding at the high speed optical core is
presented in this work.
xi
Table of Contents:
Table of Contents:....................................................................................................... xii
Chapter 1....................................................................................................................... 1
Introduction................................................................................................................... 1
1.1 Motivation and Problem definition..................................................................... 1
1.2 Thesis Outline ..................................................................................................... 2
References................................................................................................................. 5
Chapter 2....................................................................................................................... 6
Optical Data Networks for Packet Traffic .................................................................... 6
2.1 Optical data networks – current state of art ........................................................ 8
2.1.1 Optical Ethernet networks............................................................................ 8
2.1.2 Synchronous optical networks – IP over SONET...................................... 12
2.2 OPS networks of today ..................................................................................... 14
2.2.1 KEOPS network......................................................................................... 15
2.2.2 All Optical Label Swapping network ........................................................ 17
2.3 Challenges in building optical packet switched networks ................................ 19
2.3.1 Framing requirements ................................................................................ 19
2.3.2 Physical and data link layer issues............................................................. 21
2.4 OPS Performance Measurement Metrics.......................................................... 22
2.4.1 Physical Layer (Layer 1) Performance Measurements.............................. 22
xii
2.4.2 Packet level (Layer 2/3) Measurements..................................................... 24
2.5 Chapter summary .............................................................................................. 25
References............................................................................................................... 26
Chapter 3..................................................................................................................... 29
Edge Node Traffic Adaptation.................................................................................... 29
3.1 Architectural Considerations for an All-Optical Label Swapped Packet
Switched network.................................................................................................... 30
3.1.1 Traffic characteristics at the Network Edge............................................... 31
3.1.2 Electronics in optical networks.................................................................. 33
3.2 Optical Framing Considerations ....................................................................... 34
3.2.1 Forwarding requirements ........................................................................... 35
3.2.2 Idlers .......................................................................................................... 36
3.2.3 Edge and Core node Clock-Data Recovery (CDR) ................................... 37
3.2.4 Framing structure ....................................................................................... 39
3.3 Edge Node Adaptation for OPS Traffic............................................................ 42
3.3.1 Ingress Node Description........................................................................... 43
3.3.2 Egress Node Description............................................................................ 45
3.4 Edge Node Implementation .............................................................................. 46
3.4.1 PoS Framer/ Deframer ............................................................................... 47
3.4.2 Electronic Control design .......................................................................... 48
3.4.3 Tunable Laser board and control ............................................................... 51
3.5 Edge Adaptation testing and performance........................................................ 53
xiii
3.5.1 Impact of optical framing on utilization .................................................... 54
3.5.2 Measured throughput performance ............................................................ 55
3.5.3 Measured latency performance .................................................................. 56
3.6 Chapter Summary ............................................................................................. 57
References............................................................................................................... 59
Chapter 4..................................................................................................................... 60
All-Optical Label Swapping Network Performance................................................... 60
4.1 All-Optical forwarding using Wavelength Conversion .................................... 61
4.1.1 Statistical Demultiplexing.......................................................................... 63
4.1.2 Core node issues in optics and electronics................................................. 64
4.2 Optical Core Node ............................................................................................ 65
4.2.1 Two-stage wavelength conversion principle ............................................. 68
4.2.2 First stage conversion and label erase........................................................ 69
4.2.3 Second stage conversion and packet steering ............................................ 72
4.2.4 Framing and idlers revisited....................................................................... 73
4.3 Experimental Performance Evaluation ............................................................. 74
4.3.1 Measurement test system and setup........................................................... 77
4.3.2 Measured throughput performance ............................................................ 79
4.3.3 Measured latency performance .................................................................. 80
4.4 Chapter Summary ............................................................................................. 81
References............................................................................................................... 82
Chapter 5..................................................................................................................... 83
xiv
Core Traffic Adaptation for Statistical Multiplexing.................................................. 83
5.1 Core node with contention resolution............................................................... 84
5.1.1 Statistical Multiplexing and contention ..................................................... 84
5.1.2 Contention handling schemes .................................................................... 85
5.2 Optical Buffering techniques ............................................................................ 86
5.2.1 Feed-forward buffering schemes ............................................................... 86
5.2.2 Feedback buffering .................................................................................... 87
5.3 Compression/Decompression for contention resolution................................... 87
5.3.1 Network-wide deployment......................................................................... 88
5.3.2 Intra-node deployment ............................................................................... 90
5.3.3 Compression/Decompression process ....................................................... 91
5.4 Chapter Summary ............................................................................................. 93
References............................................................................................................... 95
Chapter 6..................................................................................................................... 96
Optical Packet Compression ....................................................................................... 96
6.1 Compression – State of the art today ................................................................ 97
6.1.1 Feed forward delay line approach.............................................................. 98
6.1.2 Spectral Slicing approach ........................................................................ 100
6.1.3 Loop based approach ............................................................................... 101
6.2 Network deployment of packet compression.................................................. 102
6.2.1 Intra-node compression/decompression................................................... 103
6.2.2 Packet compression classification............................................................ 106
xv
6.2.3 Requirement of a practical compressor.................................................... 107
6.2.4 Approaches to compression implementation ........................................... 109
6.3 Fold-in packet compression ............................................................................ 111
6.3.1 Linear time delay compression ................................................................ 112
6.3.2 Loop based compression.......................................................................... 113
6.4 Packet adaptation using fold-in compression ................................................. 115
6.4.1 Loop characterization............................................................................... 116
6.4.2 2.5Gbps to 10Gbps Packet compression.................................................. 117
6.4.3 Variable compression control .................................................................. 121
6.4.4 Bandwidth efficient multiple compressor approach ................................ 124
6.5 Chapter summary ............................................................................................ 126
References............................................................................................................. 129
Chapter 7................................................................................................................... 131
Optical Packet Decompression ................................................................................. 131
7.1 Decompression – State of the art .................................................................... 132
7.1.1 Feed-forward Delay line approach........................................................... 133
7.1.2 Spectral slicing......................................................................................... 135
7.1.3 Loop based approach ............................................................................... 136
7.2 Fundamentals, design and implementations of packet decompression .......... 137
7.2.1 Decompression - principles of operation ................................................. 137
7.2.2 Requirements on a packet decompressor................................................. 139
7.3 Fold-out packet decompression ...................................................................... 140
xvi
7.3.1 Linear time delay approach...................................................................... 141
7.3.2 Loop based approach ............................................................................... 142
7.3.3 Loop characterization............................................................................... 144
7.3.4 10Gbps to 2.5Gbps packet decompression .............................................. 146
7.3.5 Variable packet decompression ............................................................... 149
7.4 Chapter summary ............................................................................................ 151
References............................................................................................................. 152
Chapter 8................................................................................................................... 153
End-to-End Core Traffic Adaptation ........................................................................ 153
8.1 End-to-end core traffic adaptation .................................................................. 153
8.2 Issues in end-to-end compression-decompression experimentation............... 154
8.3 10Gbps core adaptation of 2.5Gbps traffic ..................................................... 158
8.4 Variable length packet adaptation................................................................... 162
8.5 Chapter summary ............................................................................................ 164
Chapter 9................................................................................................................... 165
Conclusion ................................................................................................................ 165
9.1 Thesis conclusions .......................................................................................... 165
9.2 Future work..................................................................................................... 169
xvii
List of Figures:
Figure 2. 1: Layered 10GigE model. PHY - Physical layer device; PMD - Physical
Medium Dependent....................................................................................................... 9
Figure 2. 2: General Ethernet Framing Format. CRC - Cyclic Redundancy Check;
MAC - Media Access Control .................................................................................... 10
Figure 2. 3: Basic SONET Frame. SPE - Synchronous Payload Envelope................ 13
Figure 2. 4: HDLC-framed PPP-encapsulated IP packet............................................ 13
Figure 2. 5: PoS OSI Stack. IP - Internet Protocol; PPP- Point to Point Protocol;
HDLC - High-level Data Link Control; SONET - Synchronous Optical NETwork.. 14
Figure 2. 6: KEOPS Packet Format ............................................................................ 15
Figure 2. 7: KEOPS Node structure. DMUX - Demultiplexer; MUX - Multiplexer . 16
Figure 2. 8: AOLS Packet Frame. (a) Serial label, (b) Optical SCM label ................ 17
Figure 2. 9: AOLS Network architecture.................................................................... 18
Figure 2. 10: IP Packet Structure. IHL - Internet Header Length; TOS - Type Of
Service; TTL - Time To Live...................................................................................... 20
Figure 2. 11: Data Entity Granularity ......................................................................... 21
Figure 3. 1: End-to-End network functional architecture. .......................................... 31
Figure 3. 2: a) Peer-to-Peer vs. (b) Overlay optical network approach ...................... 32
Figure 3. 3: Need for idlers - transmission distortions in the absence of dc balance.. 37
xviii
Figure 3. 4: Packet mode clock recovery.................................................................... 38
Figure 3. 5: Optical Framing Structure. SW - Synch. Word; S/EOOP –Start/End Of
Optical Packet; OL - Optical Label; S/EOIP- Start/End Of IP Packet ...................... 40
Figure 3. 6: Example 32 bit Optical Label or header. O_ID - Optical Identifier; TTL -
Time To Live .............................................................................................................. 41
Figure 3. 7: Experimental scope traces of (a) Optically framed packets (b) Packet
framing (refer Figure 3.6) ........................................................................................... 42
Figure 3. 8: Edge node adaptation layering ................................................................ 43
Figure 3. 9: Ingress node high level schematic........................................................... 44
Figure 3. 10: Egress node functionality. CDR- Clock Data Recovery ....................... 45
Figure 3. 11: Vitesse POS OC-48 Framer/Deframer functionality............................. 47
Figure 3. 12: Ingress FPGA process functional diagram............................................ 49
Figure 3. 13: Egress FPGA process functionality....................................................... 50
Figure 3. 14: Egress FPGA - Phase locker operating principle .................................. 51
Figure 3. 15: Fast tunable laser functionality. DAC – Digital to Analog Converter; V/I
– Voltage to Current converter ................................................................................... 52
Figure 3. 16: Fast tunable laser spectrum at 1549.32nm and switching scope trace .. 52
Figure 3. 17: Test path and traffic flow setup............................................................. 53
Figure 3. 18: Edge adaptation theoretical throughput calculation .............................. 54
Figure 3. 19: Edge adaptation throughput measurement and theoretical comparison 55
Figure 3. 20: Edge adaptation latency measurement .................................................. 57
xix
Figure 4. 1: 4 x 4 AWGR wavelength mapping - passive, non-blocking, static optical
switch; λji – input port i, wavelength j ................................................................. 62
Figure 4. 2: Statistical demultiplexing (a) 1 input (b) 4 inputs - forwarding without
contention resolution................................................................................................... 63
Figure 4. 3: Optical Core Node - Principle of operation............................................. 65
Figure 4. 4: Core FPGA Process - 4 control signals. OL – Optical Label; S/EOOP –
Start/End Of Optical Packet; DMUX – Demultiplexer. ............................................. 67
Figure 4. 5: 2-Stage wavelength conversion principle. BPF - Band Pass Filter; SOA -
Semiconductor Optical Amplifier; MZI - Mach Zehnder Interferometer .................. 69
Figure 4. 6: SOA based 1st stage Wavelength Conversion and frame erase.............. 69
Figure 4. 7: Op Amp based Fast switching current driver circuitry for SOA............. 70
Figure 4. 8: Output power of the SOA as a function of the input voltage applied ..... 71
Figure 4. 9: Optical Signal to Noise Ratio (OSNR) performance of first stage
wavelength converter system...................................................................................... 72
Figure 4. 10: Mach Zehnder Interferometer based 2nd stage Wavelength Conversion.
FBG – Fiber Bragg Grating ........................................................................................ 73
Figure 4. 11: External Idler Insertion. AWGR - Arrayed WaveGuide Router ........... 74
Figure 4. 12: Core node experimental setup ............................................................... 75
Figure 4. 13: 1st stage Wavelength Converter Scope traces (a) CW + packet (b)
Frame erased wavelength converted payload ............................................................ 76
Figure 4. 14: Scope traces after 2nd stage Wavelength Conversion (a) Before rewrite
(b) After frame rewrite................................................................................................ 77
xx
Figure 4. 15: Traffic flow and Layer-3 measurement setup ....................................... 78
Figure 4. 16: End-to-End IP throughput measurement and theoretical comparison
(refer Figure 4.15)....................................................................................................... 79
Figure 4. 17: End-to-End latency vs. packet size (refer Figure 4.15)......................... 80
Figure 5. 1: Contention resolution during statistical MUX. CR- Contention Resolution
..................................................................................................................................... 84
Figure 5. 2: Feed-forward delay structure................................................................... 86
Figure 5. 3: Feedback delay buffering ........................................................................ 87
Figure 5. 4: Compression/Decompression deployed at the network edges ................ 89
Figure 5. 5: % available link utilization Network scale ComDecom vs. packet size in
bytes ............................................................................................................................ 90
Figure 5. 6: Intra-node deployment of Compression/decompression at the core ....... 91
Figure 5. 7: Compression Decompression process detail ........................................... 92
Figure 5. 8: RZ pulse quality requirements for compression...................................... 93
Figure 6. 1: (a) Structure of a feed forward delay compressor, (b) Working Principle
..................................................................................................................................... 99
Figure 6. 2: Spectral slicing approach - Principle of operation ................................ 100
Figure 6. 3: Loop based compression - Principle of operation................................. 101
Figure 6. 4: Intra node deployment of compression and decompression ................. 103
Figure 6. 5: Edge ingress node functionality– electronic adaptation........................ 104
xxi
Figure 6. 6: Edge egress functionality – electronic adaptation................................. 104
Figure 6. 7: Intra-node deployment of Compression/decompression at the core ..... 105
Figure 6. 8: Fixed compression ratio compression – padding to achieve fixed
temporal size ............................................................................................................. 107
Figure 6. 9: Variable and discretely variable compression ratio compression ......... 107
Figure 6. 10: Generic structure of compressor/decompressor .................................. 108
Figure 6. 11: Bit- serial compression........................................................................ 110
Figure 6. 12: Compression Principle - packet fold-in approach ............................... 111
Figure 6. 13: Linear time delay implementation of fold-in compression ................. 112
Figure 6. 14: Loop based compression principle ...................................................... 114
Figure 6. 15: Loop based implementation of fold-in compression (colors indicate
virtual bins not wavelengths) .................................................................................... 115
Figure 6. 16: Average Power increase at the input of the loop SOA........................ 116
Figure 6. 17: Measured SOA power penalty, OSNR Vs input power ...................... 117
Figure 6. 18: 2.5Gbps to 10Gbps compression - Experimental setup ...................... 118
Figure 6. 19: (a) Input 1500 packet and (b) packet stream quality .......................... 119
Figure 6. 20: (a) Compressed 1500 byte packet and (b) compressed stream quality
................................................................................................................................... 120
Figure 6. 21: Bit error rate measurements on 1500 byte compressed packets (refer
Figure 6. 18).............................................................................................................. 121
Figure 6. 22: Principle of variable compression ratio control .................................. 122
xxii
Figure 6. 23: 1500, 1024 byte input packets (a1,b1) and compressed output packets
(a2,b2) ....................................................................................................................... 122
Figure 6. 24: 560, 40 byte input packets (c1,d1) and compressed output packets
(c2,d2) ....................................................................................................................... 124
Figure 6. 25: Bandwidth efficient compression setup. ............................................. 126
Figure 7. 1: Feed forward delay decompression- (a) structure (b) operation ........... 133
Figure 7. 2: Decompression using spectral slicing ................................................... 135
Figure 7. 3: Loop based decompression structure and operation.............................. 136
Figure 7. 4 : Decompression principle - increase in utilization ................................ 138
Figure 7. 5: Decompression - Time out issue ........................................................... 139
Figure 7. 6:Linear time delay decompression technique .......................................... 142
Figure 7. 7:Loop based decompression principle .................................................... 143
Figure 7. 8: Loop based fold-out decompression...................................................... 144
Figure 7. 9: Decompression SOA power penalty, OSNR, output power Vs input
power......................................................................................................................... 145
Figure 7. 10: Loop based decompression - experimental setup................................ 146
Figure 7. 11: 10Gbps to 2.5 Gbps decompression of 1500 byte packets - (a1,b1) input
compressed packets and bit quality, (a2,b2) decompressed packets and bit quality
(2μsec (left) and 100psec (right) time scale) ............................................................ 147
Figure 7. 12: Bit error rate measurement on decompressed 1500 byte packets (refer
Figure 7. 10).............................................................................................................. 149
xxiii
Figure 7. 13: Compressed packet input(a1, b1, c1, d1) and decompressed packet
output (a2, b2, c2, d2) for 1500, 1024, 560 and 40 byte packets ( 500nsec time scale)
................................................................................................................................... 150
Figure 8. 1: Loop based fold-in compressor output imperfection ............................ 155
Figure 8. 2: Loop based fold-out decompressor output imperfection....................... 156
Figure 8. 3: (De)encoding to enable compression/decompression ........................... 156
Figure 8. 4: Compression and decompression of encoded packets to prevent payload
bit corruption............................................................................................................. 157
Figure 8. 5: End -to-end compression/decompression experimental setup .............. 159
Figure 8. 6: 1500 bytes (a1, a2) Input packet ,bit quality at 2.5Gbps (b1, b2)
compressed packet, bit quality at 10Gbps................................................................ 160
Figure 8. 7: 1500 byte packets (a1, a2) compressed packet and bit quality at 10Gbps,
(b1, b2) decompressed packet and bit quality at 2.5Gbps ........................................ 161
Figure 8. 8: End-to-End bit error rate measurement at 2.5Gbps............................... 162
Figure 8. 9: Input, compressed and decompressed packets for (a) 1500 byte, (b) 1024
byte, (c) 560 byte, (d) 40 byte packets...................................................................... 163
xxiv
Chapter 1
Introduction
1.1 Motivation and Problem definition
Mankind’s passion for interaction and exchange of information has fueled his ability
to communicate from just a handful of primal sounds to using packets of light to get
the message across. In today’s day and age, emails and web pages containing millions
of bits of information are routinely created and read everyday to the extent that they
have now become more of a necessity than a luxury. The Internet penetration has
grown from 0.4% in December 1995 to 15.7% of the world population in December
2005 [1]. Internet growth is reported to be about 100% every 12-18 months [2] and
newer avenues of internet usage such as video broadcast over the internet and IP
telephony are growing in popularity. With the ever increasing demand for network
bandwidth and higher degree of service sophistication required such as Quality Of
Service (QOS), hardware supporting the internet namely electronic routers are close
to reaching their physical limits [3].
Optical packet switching provides a scalable alternative to forwarding traffic in the
electronic domain. Advantages in switching in optics include lower power
consumption, smaller physical footprint, and it’s potential to scale both in bit rate and
1
packet forwarding rate [4]. Several key technologies in processing data in optics have
been researched over the past few years and the maximum bit rate handled in optics
has steadily increased. However, research in bridging the gap between increasing IP
traffic in today’s internet and enabling optical technologies has been limited. Another
critical issue in forwarding packets in optics is contention resolution. While the IP
traffic characteristics dictate variable packet sizes arriving randomly, core optical
switches with limited buffering capabilities are better suited to handling data in fixed
time slots. The purpose of the research done in this thesis is to investigate an
appropriate opto-electronic adaptation mechanism that enables both transport and
forwarding of raw IP packets taking into account their current traffic characteristics
and using state of the art all-optical technologies.
1.2 Thesis Outline
In chapter 2, we examine current optical transport protocols and some of the
significant state of the art research in optical packet switching with the aim of
identifying known issues in building an All-Optical Packet Switched network. We
take a look at different performance metrics required both in the physical layer testing
of optical sub-systems and network layer testing for end-to-end IP packet forwarding
that will be used in subsequent chapters.
Edge node packet adaptation requirements are investigated and an optical framing
mechanism that takes all known issues into account and satisfies these adaptation
requirements is proposed in Chapter 3. Ingress and egress functionality are studied
2
and individual sub-systems that are required for implementation of adaptation from IP
traffic to and from optical packet traffic are described. Theoretical and experimental
results on end-to-end packet performance through the edge adaptation are presented
and the impact of optical framing is investigated.
Chapter 4 deals with the basic principle behind forwarding traffic at the optical core
node. Statistical demultiplexing using 2-stage wavelength conversion as an all-optical
packet steering mechanism is studied. The electronic and optical technologies
required in enabling all-optical packet forwarding are presented along with the
identification of bottlenecks in packet forwarding at the core. An experimental
demonstration of all-optical packet forwarding made possible due to the edge node
adaptation is shown and the impact of core node packet forwarding on layer-3 end-to-
end network performance is analyzed.
In chapter 5, we look at the requirements of the optical core node to perform
statistical multiplexing of packet traffic. Current solutions to packet contentions are
examined and the issues with optical buffering technology are identified.
Compression and decompression, as an optical adaptation mechanism to adapt
variable sized payloads to fixed time slots desired in optical buffering is proposed and
deployment strategies for such adaptation within the network are compared.
Current state of the art in optical packet compression is reviewed in Chapter 6 and the
problems in implementing various schemes to compress IP packets are discussed.
Requirements for a practical compressor are summarized based on knowledge of
input and the desired output traffic characteristics. We then propose a novel approach
3
to packet compression that is capable of handling large packet byte sizes and
investigate two techniques of implementing this approach. The sub-systems that go
into the practical realization of packet compression using the one proposed approach
namely loop-based compression are examined and we present an experimental
demonstration of packet compression capable of compression using variable
compression ratios to achieve variable length to fixed packet size adaptation. Finally,
the performance of the compression scheme is evaluated using physical layer metrics.
Our research on optical packet decompression is presented in Chapter 7 along with
prior state of the art in this area. We put forward an approach to decompression that
complements the compression scheme presented in chapter 6 and the feasibility of
such an approach is experimentally demonstrated. Bit error rate measurements are
presented to evaluate the performance of such a decompression scheme. We look at
the issues in implementing back-to-back compression and decompression and
propose modifications to edge node adaptation to solve the identified issues. Finally
an experimental demonstration of end-to-end packet compression and decompression
for variable packet sizes as an adaptation to a fixed time slot is presented and its
performance is evaluated.
Chapter 8 presents the thesis conclusions and possible avenues for future work in this
area.
4
References
[1] Online : http://www.internetworldstats.com/emarketing.htm.
[2] Online: http://www.dtc.umn.edu/~odlyzko/doc/itcom.internet.growth.pdf.
[3] J. Gripp, M. Duelk, J. E. Simsarian, A. Bhardwaj, P. Bernasconi, O. Laznicka,
and M. Zirngibl, "Optical switch fabrics for ultra-high-capacity IP routers,"
Lightwave Technology, Journal of, vol. 21, pp. 2839-2850, 2003.
[4] D. J. Blumenthal, "Optical packet switching,"910-912 Vol.2, 2004.
5
Chapter 2
Optical Data Networks for Packet
Traffic
Today’s communication networks need to meet several critical requirements such as
providing scalable network bandwidth, a fine granularity of control, Quality Of
Service (QOS) and backward compatibility with a wide range of existing network
protocols and framing formats. Fiber optic communication has been used for years as
a transport medium in data networks for rapid transfer of large amounts of traffic
between nodes. However, at each processing node, data traffic gets converted to
electronic form before examination and reconverted to optical form after processing
and before exiting the node. Moore’s law predicts the growth of future electronic
circuitry and states that the number of transistors on an Integrated Circuit (IC)
doubles every 18-24 months. However, it has been seen that with the birth of the
internet, data traffic far exceeds growth predictions for electronic processing
capability made by Moore’s law. Consequently, it can be deduced that in view of the
explosive nature of data traffic, the direct processing of traffic using electronics
would require parallel electronic processors to meet future forwarding requirements.
However, such a scaling mechanism is cumbersome and non-optimal. Minimizing the
6
need for O-E-O conversion within the network is one approach to alleviate the burden
on the electronic domain. The aim of using Optical Packet Switching (OPS) is to
separate forwarding decisions that may be made at packet rates by existing
electronics, from the ultra-fast data rates of the packet payloads supported by fiber
optics, while still maintaining a fine granularity of control. Significant research is
being done targeting minimization of power and space footprints of optical
components by integration of multiple functionalities into a single chip [1, 2].
However, several issues need to be solved before deploying such chips to improve the
network performance. The purpose of the research presented in this thesis is
threefold:
• To explore the adaptation requirements on traffic within an OPS network
• To minimize bottlenecks on handling packet traffic due to practical
limitations of optical forwarding and the electronic control plane and
• To propose, demonstrate and evaluate optical adaptation techniques with the
intent of solving buffering issues that can be implemented using today’s
optical technologies.
In this chapter, we briefly examine today’s optical networks and extract requirements
that would need to be met when building an All-Optical Packet Switched (AOPS)
network capable of forwarding IP traffic.
7
2.1 Optical data networks – current state of art
Optical data networks in place today are primarily used as a transport mechanism
geared towards providing an efficient communication medium while a higher layer
(IP or ATM) implemented in electronics performs switching of individual routable
units of data. Such optical networks are designed to rely heavily upon the electronic
domain for all forwarding operations and consequently requiring as many as six O-E-
O conversions at each route processing node [3]. In order to grow successfully, data
networks must satisfy very different requirements and cater to end-user IP networks
and Corporate Storage Area Networks (SAN) [4]. In this section we examine two of
the most popular optical data networks currently in place namely Optical Ethernet
(Gigabit Ethernet or Gig-E and 10 Gigabit Ethernet or 10GigE) networks and the
Synchronous Optical NETwork (SONET) networks, to understand the reasons behind
their widespread deployment and identify the constraints and drawbacks of today’s
optical networks.
2.1.1 Optical Ethernet networks
While copper based Ethernet traditionally dominated Local Area Networks (LAN)
with 95% of IP traffic originating and terminating in Ethernet [5], optical Gig-E
networks which consists of 1000Mbps and 10Gbps links are designed to be deployed
as backbones in Metro Area Networks (MAN) and more recently Wide Area
Networks (WAN) [6]. Ethernet was originally developed by Robert Metcalfe and
David Boggs at the Xerox Corporation in 1973 to work at 2.94Mbps. All Ethernet
8
networks follow the Carrier Sense Multiple Access with Collision Detect
(CSMA/CD) protocol [ref IEEE 802.3] originally intended for the half duplex
operation, where multiple stations shared a single communication medium. The key
to success in ethernet is its simplicity and the high degree of structured compatibility
offered over various physical and data link implementations. While Gigabit ethernet
supports copper and optical transmission, 10GigE addresses many concerns arising
during high speed transport of packet traffic. We therefore review 10GigE standards
in particular as defined in [ref. IEEE standards document 802.3ae]. The most
important difference between 10GigE and previous Ethernet flavors is the elimination
of the half duplex operation or in essence the CSMA/CD and using only full duplex
operation, as half-duplex operation becomes cumbersome at high speeds. Figure 2. 1
shows the general layered model of a 10GigE network.
Session
Presentation
Application
Transport
Network
Data Link
Physical
Logical Link Control (LLC) and Media Access Control (MAC)
Full Duplex
ReconciliationReconciliation
XGMII (and XGAUI)
WWDM PHYWAN PHYLAN PHY
PMD PMD PMD PMD PMDPMD PMD
Session
Presentation
Application
Transport
Network
Data Link
Physical
SessionSession
PresentationPresentation
ApplicationApplication
TransportTransport
NetworkNetwork
Data LinkData Link
PhysicalPhysical
Logical Link Control (LLC) and Media Access Control (MAC)
Full Duplex
Logical Link Control (LLC) and Media Access Control (MAC)
Full Duplex
ReconciliationReconciliationReconciliationReconciliation
XGMII (and XGAUI)XGMII (and XGAUI)
WWDM PHYWWDM PHYWAN PHYWAN PHYLAN PHYLAN PHY
PMDPMD PMDPMD PMDPMD PMDPMD PMDPMDPMDPMD PMDPMD
Figure 2. 1: Layered 10GigE model. PHY - Physical layer device; PMD - Physical Medium Dependent
As can been seen, the strongly structured Ethernet model has a Media Independent
Interface (MII) along with a physical layer definition specific to a variety of
implementations and is an important reason behind the widespread deployment of
9
Ethernet. Key issues defined in Ethernet standards that are relevant to any optical
adaptation are encoding, physical layer clock/data recovery mechanisms, framing
specifications and reconciliation between transmission rates and electrical processor
rates. Data from higher layers is passed on to a Logical Link Control (LLC) and
Media Access Control (MAC), which are responsible for framing the data and control
handshaking for successful transfer of frames. 10GigE specifies a 10 Gigabit Media
Independent Interface (XGMII) and reconciliation layer, which provide a simple,
inexpensive, and easy to implement interface between the MAC and the lower
Physical Layer Device (PHY) layers. In essence, one function of these layers is to
handle proper serialization and deserialization of data between the physical layer
serial 10Gbps and MAC layer parallel 155 Mbps data rates. 10GigE defines seven
unique physical layer interfaces, 6 serial and 1 Wide Wavelength Division
Multiplexed (WWDM) mainly to cater to a wide range of cost, range (for LANs and
WANs) and compatibility with other technologies such as SONET. Due to its low
latency and high speed, 10GigE has become promising for Storage Area Networks
(SANs) and several companies such as Cisco, Extreme networks and Foundry
networks offer 10GigE interfaces today. The basic framing format used in all
Ethernet networks is the same and is shown in Figure 2. 2.
Preamble8 bytes
D Addr6 bytes
S Addr6 bytes
Type2 bytes
Data46 – 1500 bytes
CRC4 bytes
MAC Header Data Field MAC Trailer
Preamble8 bytes
D Addr6 bytes
S Addr6 bytes
Type2 bytes
Data46 – 1500 bytes
CRC4 bytes
MAC Header Data Field MAC Trailer
Figure 2. 2: General Ethernet Framing Format. CRC - Cyclic Redundancy Check; MAC - Media
10
Access Control
Bits are considered only in octets and frames are separated by a minimum interframe
gap of 12 bytes for synchronization at various layers. Small packets are padded up to
the minimum packet size of 64 bytes and the maximum supported frame size is
1518bytes. Each Ethernet frame consists of a start byte, 7 preamble bytes, a source
and destination address of 6 bytes each, the data of 64 - 1518 bytes and a trailer
Cyclic Redundancy Check (CRC) of 4 bytes. In total, 10GigE has a 26 byte overhead
per packet and a 12 byte minimum interpacket gap requirement or a 38 byte aggregate
overhead per packet amounting to 97.5 % utilization for 1500 byte packets. 10GigE
uses continuous signaling between the transmitter and the receiver so that clock may
be recovered from the data channel. Clock recovery is therefore not performed on a
frame-by-frame basis as was the case in early Ethernet implementations. A 64B/66B
encoding scheme is used by 10GigE, replacing the less efficient 8B/10B scheme in
earlier Ethernet implementations to maintain DC balance, increase data bit transitions
for clock recovery and increase the possibility of error detection and correction.
10GigE therefore offers a very standardized framing structure that may be
used with a wide variety of physical interfaces with the possibility of tapping into
cheap and efficient bandwidth. However, any forwarding decision will have to be
made only after the entire packet has been examined and therefore is wasteful in
terms of electronic processing requirement.
11
2.1.2 Synchronous optical networks – IP over SONET
Another widely deployed standard for optical communication in today’s
communication networks is Synchronous Optical NETwork (SONET). It is an
interface standard for connecting fiber optic transmission systems and fits in the
physical layer of the Open System Interconnect (OSI) model. SONET was originally
proposed by Bellcore in the mid 80s and is currently well defined by the ANSI [7].
The standard was originally created to run at OC-1 (51.84Mbps) with the goal of
unifying international digital systems and as a way to multiplex digital channels to
higher data rates of OC-192 (9.95Gbps) or higher. While SONET is strictly a physical
layer standard, it lends itself to transport IP directly over it as Packet over SONET
(PoS).
The basic SONET frame is shown in Figure 2. 3 consists of 810 bytes of information
that may be viewed as 90 bytes wide and 9 rows high [8]. The user data or
Synchronous Payload Envelope (SPE) occupies 87 columns while 3 columns are
reserved for section overheads. Each SONET frame is transmitted at precise 125usec
intervals regardless of the data rate. Multiplexing of several SONET streams is done
on a byte-by-byte basis to achieve higher data rates of OC-3, OC-12, OC-48 etc.
12
……….9 rows
80 columns
Section
Line
Transport Overhead
SPE Path Overhead
125 usec
……….9 rows
80 columns
Section
Line
Transport Overhead
SPE Path Overhead
125 usec
Figure 2. 3: Basic SONET Frame. SPE - Synchronous Payload Envelope
All SONET nodes are synchronous or plesionchronous to reduce or eliminate the
need for bit stuffing to compensate for clock variations in streams from different
transmitters. When transmitting IP packets over SONET, each IP datagram is first
encapsulated in a Point-to-Point Protocol (PPP) [ref. RFC 1661] and then in framed
using High-level Data Link Control (HDLC) [ref. RFC 1662] before being scrambled
and fit inside the SPE. Figure 2. 4 shows the general format of a HDLC-framed PPP-
encapsulated IP datagram.
Flag1 byte
Address1 byte
Protocol2 bytes
DataUsually 46 – 1500 bytes
FCS4 bytes
Flag1 byte
Flag1 byte
Address1 byte
Protocol2 bytes
DataUsually 46 – 1500 bytes
FCS4 bytes
Flag1 byte
Figure 2. 4: HDLC-framed PPP-encapsulated IP packet
The PPP encapsulation provides multiprotocol encapsulation, error control and link
initialization control while the HDLC framing is mainly used to provide delineation
within the SONET frames. SONET uses 1+x43 scrambling to ensure sufficient bit
transitioning and for security. The OSI protocol stack for PoS is shown in Figure 2. 5.
PoS uses a minimum of 9 bytes to encapsulate each IP packet and 36 bytes for every
13
810 bytes as SONET overhead. Framing multiple IP packets into synchronous
containers reduces the total framing overhead and the effective utilization for 1500
byte packets may be as high as 99.5%.
IPPPP
HDLCSONET
Network Layer
Data Link Layer
Physical Layer
IPPPP
HDLCSONET
Network Layer
Data Link Layer
Physical Layer
Figure 2. 5: PoS OSI Stack. IP - Internet Protocol; PPP- Point to Point Protocol; HDLC - High-level Data Link Control; SONET - Synchronous Optical NETwork
While PoS is an efficient way of transferring IP packets, it does not allow for
individual manipulation of packets during forwarding. All framing must be removed
before IP packets can undergo statistical multiplexing/demultiplexing and the
forwarding electronics must therefore examine the entire data stream to do so.
2.2 OPS networks of today
While the static provisioning of optical fiber bandwidth has been widely used
commercially using Wavelength Division Multiplexing (WDM) in fibers, OPS
affords the possibility of more efficiently utilizing available bandwidth by
provisioning it dynamically at a much finer granularity. Optical technologies capable
of supporting OPS have been demonstrated and several network approaches such as
KEOPS, AOLS, WASPNet and OPSNet have been proposed. In this section we will
briefly look at two of the significant proposals for OPS and define key issues that
need to be addressed to enable packet forwarding in the optical domain.
14
2.2.1 KEOPS network
Work on the European ACTS approach to OPS has been done under the KEys to
Optical Packet Switching (KEOPS) project [9] with contributions from both
companies such as Alcatel and France Telecom and universities such as TUD and
ETH – Zurich. The focus here was to analyze and address some basic issues before
optical bit rate transparent packet switching can be performed. The approach proposes
to use fixed duration payloads that are 1.35usecs long with both forwarding
information and the payload being transmitted serially on the same wavelength.
Figure 2. 6 shows the general packet structure used while Figure 2. 7 shows the basic
node structure used in the project. An important issue with this approach is that the
payload time must be adjusted closely relative to the header using synchronizers.
Incoming packets undergo O-E conversion for header recovery and synchronization
Synch
Header
Payload 1.35usecs
1.646 usec packet
Guard times time
Synch
Header
Payload 1.35usecs
1.646 usec packet
Guard times time
Figure 2. 6: KEOPS Packet Format
before entering the switch fabric. Within the switch, packets undergo header erasure,
demultiplexing, wavelength conversion, multiplexing and finally header rewriting
before exiting. Mach Zehnder Interferometric Wavelength Converters (MZ – IWC)
15
are used to tune packets to appropriate wavelengths. This approach uses input
buffering [10] in the switch to avoid collisions.
Header Recovery and synch control
Switch Control Header Update
Electrical Control Plane
Synch
Switch Fabric
Signal regeneration and header
updation
DMUX MUX
Optical Plane
Fiber in Fiber out
Header Recovery and synch control
Switch Control Header Update
Electrical Control Plane
Synch
Switch Fabric
Signal regeneration and header
updation
DMUX MUX
Optical Plane
Fiber in Fiber out
Figure 2. 7: KEOPS Node structure. DMUX - Demultiplexer; MUX - Multiplexer
However a feed-forward approach to buffering is adopted which does not provide
good support for differentiated services. While the project addresses some key
concerns of OPS and demonstrates essential enabling optical components, actual end-
to-end demonstration of packet forwarding is not undertaken. The use of a fixed
payload size requires aggregation of smaller IP packet sizes at the edge for efficient
bandwidth utilization. This can have undesirable effects on the overall latency and
places further processing requirements on the edge electronics. Packets longer than
the fixed payload size would require segmentation before being forwarded through
the optical network. Segment misordering and latency increase due to additional
processing steps are some of the drawbacks in doing so. Moreover, while packet
16
header may be used for identification of the optical packet itself at core nodes [11],
the need for payload encapsulation is not addressed.
2.2.2 All Optical Label Swapping network
Another approach to OPS is the All-Optical Label Swapping (AOLS) done at the
University of California, Santa Barbara (UCSB) [12]. Work done in this thesis adopts
the AOLS architecture and investigates the adaptation requirements in implementing
this architecture. The approach considers both serial and Sub-Carrier Multiplexing
(SCM) approaches for labeling of optical packets as shown in Figure 2. 8., introduces
specific technologies required to implement the architecture and finally demonstrates
high speed label swapping using fiber based wavelength converters (FWC).
Variable sized IP Payload IP Header Optical Label
Guard Band
IP Bit Rate e.g., 2.5 Gbps – 40 Gbps Optical Label Bit Rate e.g. 2.5Gbps
(a)
Variable sized IP Payload IP Header
Optical Label
IP Bit Rate e.g., 2.5 Gbps – 40 Gbps
Optical Label Bit Rate e.g. 2.5Gbps
Optical SCM Label
Encapsulation
Optical Serial Label
(b)
Variable sized IP Payload IP Header Optical Label
Guard Band
IP Bit Rate e.g., 2.5 Gbps – 40 Gbps Optical Label Bit Rate e.g. 2.5Gbps
(a)
Variable sized IP Payload IP Header
Optical Label
IP Bit Rate e.g., 2.5 Gbps – 40 Gbps
Optical Label Bit Rate e.g. 2.5Gbps
Optical SCM Label
Encapsulation
Optical Serial Label
(b)
Figure 2. 8: AOLS Packet Frame. (a) Serial label, (b) Optical SCM label
17
Figure 2. 9. shows the general AOLS architecture where IP packets are labeled at the
ingress nodes, forwarded all optically at the core node and finally stripped back to
their IP framing before exiting the network at the egress nodes. Fiber Loop Mirrors
(FLM) used for SCM header recovery, fast switching tunable laser sources, SOA
based MZ-IWCs and high speed FWCs have been demonstrated. While the general
framework for OPS has been researched, specific adaptation and actual demonstration
of end-to-end forwarding of IP packets remains to be done.
Ingress Edge Router with
AOLS Interface
AOLS core router or subnet
Egress Edge Router with
AOLS InterfaceAOLS core router or subnet
IP Packet from source node Optical
Label and packet at λi Optical
Label and packet at λj
IP Packet to destination node
Optical AOLS Core Network
Ingress Edge Router with
AOLS Interface
Ingress Edge Router with
AOLS Interface
AOLS core router or subnet
AOLS core router or subnet
Egress Edge Router with
AOLS Interface
Egress Edge Router with
AOLS InterfaceAOLS core router or subnet
AOLS core router or subnet
IP Packet from source node Optical
Label and packet at λi Optical
Label and packet at λj
IP Packet to destination node
Optical AOLS Core Network
Figure 2. 9: AOLS Network architecture
Research on electronic adaptation ensuring proper transport and forwarding of
individual packets and optical adaptation that may be used to improve optical
forwarding performance by addressing issues such as buffering for contention
resolution would help in the complete end-to-end realization of the AOLS
architecture.
18
2.3 Challenges in building optical packet switched networks
As described in the previous section, significant research in OPS has been done over
the past few years including but not limited to research in architectural designs, sub-
systems development and component integration [1, 2]. Consequently, OPS has
reached a state of maturity where end-to-end network design, demonstration and
performance assessment for packet forwarding may be attempted. However,
sufficient design of both optical and electronic adaptation would be required, taking
into account all the physical layer limitations and forwarding plane requirements. In
this section, we look at the impact of the different optical and electronic layer
constraints on optical packet framing before putting down a set of guidelines for
design of an adaptation mechanism.
2.3.1 Framing requirements
Before proceeding to define the framing requirements, we examine the nature of the
data to be forwarded through the all-optical network. While most or all of the
adaptation proposed and demonstrated in this thesis may be modified to fit different
packet framing standards external to the all-optical network, we concentrate on
forwarding Internet Protocol (IP) packets as they make up most of the today’s
Internet traffic. Figure 2. 10 shows the general structure of an IP packet. The IP
header consists of multiple fields of which the Type of Service (TOS), total length,
Time To Live (TTL) and source and destination addresses are most relevant for
forwarding in the optical domain. While IPv4 can support up to 65536 bytes of
19
payload, Internet statistics [13] show that majority of the internet traffic has an IP
packet size variation from 44 to 1500 bytes. In this thesis, incoming packets are
assumed to be of variable byte size from 44 to 1500 bytes. Forwarding of packets is
done primarily using the IP destination address, which is 32 bits long. This
information must be passed on to the optical header for proper forwarding and routing
of optical packets.
Data – variable length
Option and Padding
Destination Address
Source Address
TTL Protocol Header Checksum
Identification Flags Fragment Offset
Total LengthVer. IHL. TOS
Data – variable length
Option and Padding
Destination Address
Source Address
TTL Protocol Header Checksum
Identification Flags Fragment Offset
Total LengthVer. IHL. TOS
Figure 2. 10: IP Packet Structure. IHL - Internet Header Length; TOS - Type Of Service; TTL - Time To Live
QOS within the optical network may be set by including information from the IP TOS
field. IP packets do not undergo any scrambling and any optical framing mechanism
implemented must, if required, have its own payload scrambling for the purpose of
both security and dc balance. It is also preferable that each IP packet be transferred as
a single self-contained routable unit without fragmentation. In the case where framing
requires fragmentation of the packets, the framing mechanism must ensure that IP
packets exiting the optical network would be whole.
20
2.3.2 Physical and data link layer issues
When designing an adaptation for OPS, it is important to take physical rise/fall time
considerations of both optics and control electronic into account. While technology to
handle payload bit rates of up to 160Gbps have been demonstrated at UCSB [14],
electronics required to process framing information and generate control signals
presents a bottleneck in packet forwarding. It is therefore ideal to limit complex bit
level manipulations of the optical packet payload to the edge node electronics while
core node electronics deal only with packet level manipulations. Stream level
manipulation (signal amplification, regeneration, static or MEMS based quasi-static
WDM routing) may be limited to the core transmission optics as shown in Figure 2.
11. Nodes within the optical network would need to communicate for control plane
updation and for data link layer management. The optical adaptation design should
provide a mechanism for such communication between nodes.
Edge nodes – Bit level manipulation
Core nodes - Packet level manipulation
Core transmission -Stream level manipulation
Edge nodes – Bit level manipulation
Core nodes - Packet level manipulation
Core transmission -Stream level manipulation
Figure 2. 11: Data Entity Granularity
21
2.4 OPS Performance Measurement Metrics
Another challenge of building an OPS network is the proper evaluation of its
performance. While the performance of optical links is currently gauged using
physical layer metrics to ensure sufficient link quality, an OPS network designed
must test both physical and network layer metrics to evaluate its performance.
Performance tests on OPS thus far have been limited to extensive physical layer
measurements and some basic link layer packet loss predictions based on physical
layer measurements. However, the impact of various electronic and optical factors on
the performance of an OPS network cannot be sufficiently measured using just these.
For example, a physical layer connection with an excellent bit error rate performance
does not necessarily translate to a low packet loss rate. We present the main
measurement metrics that may be used to assess and improve the network
performance.
2.4.1 Physical Layer (Layer 1) Performance Measurements
Several physical layer performance measurements have been used in this thesis to
assess and predict the performance of the optical subsystems such as it bit error rate,
extinction ratio and Optical Signal to Noise Ratio. Additional measurements such as
bit jitter may be applicable based on the nature of the application.
2.4.1.1 Bit Error Rate:
This measurement is defined as the probability of incorrect identification of a bit by
the decision circuit of the receiver [15]. It is given in terms of a ratio and may be
22
calculated as the number of erroneous bits received divided by the total number of
bits transmitted, received, or processed over some stipulated period of time. The BER
is usually expressed as a coefficient and a power of 10; for example, 2.5 erroneous
bits out of 100,000 bits transmitted would be 2.5 out of 105 or 2.5 x 10-5. For a BER
measurement to be accurate, a sufficient number of errors must be detected that
statistically validates the measurement.
2.4.1.2 Extinction Ratio
Extinction ratio is defined as the ratio of two optical power levels of the intensity
modulated representation of a digital signal generated by an optical source, e.g., a
laser diode, where P 1 is the optical power level generated when the light source is
"on," and P 0 is the power level generated when the light source is "off."
0
1
PPRatioExtinction =
It is most commonly expressed in decibels (dB) in the logarithmic scale.
2.4.1.3 Optical Signal to Noise Ratio
The Optical Signal to Noise Ratio (OSNR), also measured in dB, is the ratio of
optical power in the optical signal and that in the noise within the signal bandwidth
(usually defined by the filter used) accompanying the signal. It is usually given in dB.
Noise
Signal
PP
RatioExtinction =
23
2.4.2 Packet level (Layer 2/3) Measurements
While physical layer measurements can be a good tool to predict the approximate
packet performance, layer 2/3 measurements are required to fully study and
understand the impact of packet conditions on the physical layer. Packet
measurements are layer 2 measurements when made with just frame identification,
and layer 3 when a more detailed check on both proper forwarding and complete
packet integrity is done before a decision on good or bad packet count is reached. In
addition to the metrics defined below, packet loss and packet jitter may also be
critical based on the type of application.
2.4.2.1 Packet Throughput
Packet throughput may be defined as the maximum packet rate at which none of the
offered packets are dropped by the device. In an OPS network, measurement of end-
to-end throughput is helpful in evaluating the limits of packet forwarding and is a
direct readout of the impact of changes in both network and physical layer
parameters.
2.4.2.2 Packet Latency
Latency may be defined as the time interval starting when the end of the first bit of
the input packet reaches the input port and ending when the start of the first bit of the
output packet is seen on the output port of the node under test. It may be used to
assess the amount of delay a packet suffers when being forwarded through a network.
24
2.5 Chapter summary
Packet transport protocols used commercially in today’s optical communication
networks and research approaches to date on Optical Packet Switching (OPS) were
examined in this chapter. While several key enabling optical technologies and
integration are being researched, various framing challenges need to be overcome
before the practical realization of OPS. Different framing formats have been proposed
and demonstrated thus far. However, in order to achieve complete decoupling of
optical payload and header information and to be able to perform any optical
adaptation that may be required at the core, a framing mechanism that supports
variable length payloads is needed. In the next chapter, we propose and evaluate the
performance of a new adaptation mechanism designed based both on lessons learnt
from previous work and additional optical and electronic constraints that may affect
the performance of all-optical packet forwarding.
25
References
[1] V. Lal, M. Masanovic, D. Wolfson, G. Fish, C. Coldren, and D. J.
Blumenthal, "Monolithic widely tunable optical packet forwarding chip in InP
for all-optical label switching with 40 Gbps payloads and 10 Gbps labels,"
presented at Optical Communication, 2005. ECOC 2005. 31st European
Conference on,25-26 vol.6, 2005.
[2] M. L. Masanovic, V. Lal, J. A. Summers, J. S. Barton, E. J. Skogen, L. G.
Rau, L. A. Coldren, and D. J. Blumenthal, "Widely tunable monolithically
integrated all-optical wavelength converters in InP," Lightwave Technology,
Journal of, vol. 23, pp. 1350-1362, 2005.
[3] D. J. Blumenthal, "Optical packet switching," presented at Lasers and Electro-
Optics Society, 2004. LEOS 2004. The 17th Annual Meeting of the IEEE
910-912 Vol.2, 2004.
[4] C. DeCusatis, "Storage area network applications," presented at Optical Fiber
Communication Conference and Exhibit, 2002. OFC 2002,443-444, 2002.
[5] M. C. Nuss, "Optical Ethernet in the Metro," presented at Lasers and Electro-
Optics Society, 2001. LEOS 2001. The 14th Annual Meeting of the IEEE,289
vol.1, 2001.
[6] J. Hurwitz and W. C. Feng, "End-to-end performance of 10-gigabit Ethernet
on commodity systems," Micro, IEEE, vol. 24, pp. 10-22, 2004.
26
[7] ANSI T.105-1991 and ANSI T1.105-1988, Digital hierarchy - optical
interface rates and formates specifications.
[8] A. S. Tanenbaum, Computer networks, 3rd ed. Upper Saddle River, N.J.:
Prentice Hall PTR, 1996.
[9] C. Guillemot, M. Renaud, P. Gambini, C. Janz, I. Andonovic, R. Bauknecht,
B. Bostica, M. Burzio, F. Callegati, M. Casoni, D. Chiaroni, F. Clerot, S. L.
Danielsen, F. Dorgeuille, A. Dupas, A. Franzen, P. B. Hansen, D. K. Hunter,
A. Kloch, R. Krahenbuhl, B. Lavigne, A. Le Corre, C. Raffaelli, M. Schilling,
J. C. Simon, and L. Zucchelli, "Transparent optical packet switching: the
European ACTS KEOPS project approach," Lightwave Technology, Journal
of, vol. 16, pp. 2117-2134, 1998.
[10] A. Pattavina, "Architectures and performance of optical packet switching
nodes for IP networks," Lightwave Technology, Journal of, vol. 23, pp. 1023-
1032, 2005.
[11] C. Guillemot, A. Lecorre, J. Kervarec, M. Henry, J. C. Simon, A. Luron, C.
Vuchener, P. Lamouler, and P. Gravey, "Optical packet switch demonstrator
assessment: packet delineation and fast wavelength routing," presented at
Integrated Optics and Optical Fibre Communications, 11th International
Conference on, and 23rd European Conference on Optical Communications
(Conf. Publ. No.: 448) 343-346 vol.3, 1997.
[12] D. J. Blumenthal, B. E. Olsson, G. Rossi, T. E. Dimmick, L. Rau, M.
Masanovic, O. Lavrova, R. Doshi, O. Jerphagnon, J. E. Bowers, V. Kaman, L.
27
A. Coldren, and J. Barton, "All-optical label swapping networks and
technologies," Lightwave Technology, Journal of, vol. 18, pp. 2058-2075,
2000.
[13] K. Thompson, G. J. Miller, and R. Wilder, "Wide-area Internet traffic patterns
and characteristics," Network, IEEE, vol. 11, pp. 10-23, 1997.
[14] W. Wei, L. G. Rau, and D. J. Blumenthal, "160 Gb/s variable length packet/10
Gb/s-label all-optical label switching with wavelength conversion and
unicast/multicast operation," Lightwave Technology, Journal of, vol. 23, pp.
211-218, 2005.
[15] G. P. Agrawal, Fiber-optic Communication Systems, Second Edition ed.
28
Chapter 3
Edge Node Traffic Adaptation
All Optical Packet Switched (AOPS) networks present an approach to forwarding
traffic where the forwarded optical payload does not undergo Optical-Electronic-
Optical (O-E-O) conversions throughout the network. However, to exploit the
advantages of such a network, a suitable adaptation that offers sufficient forwarding
control of the optical payloads using electronic control signals is required. In the
previous chapter we examined basic issues in adapting traffic such that all-optical
packet forwarding may take place at the core. Several factors contribute to the end-to-
end efficiency of an OPS network. In this chapter, we analyze the traffic
characteristics along with electronic and optical performance limitations, propose a
framing format for packet forwarding within the optical network and study the
various subsystems that go into building edge nodes to implement packet adaptation.
Finally, we evaluate the performance of edge adaptation implemented based on layer-
3 performance metrics and propose ways of increasing overall forwarding efficiency.
29
3.1 Architectural Considerations for an All-Optical Label
Swapped Packet Switched network
In implementing an OPS network, a successful packet adaptation would need to take
several influencing elements into account. At the network edge, packets transition
between the electronic and optical planes and the bulk of packet processing takes
place in the electronic domain. Consequently bit level processing of the packet stream
is possible at base rate. In order to exploit the high bandwidth capabilities offered by
optics, it is desirable that optical payloads are transmitted at higher bit rates (10Gbps
or higher) while the framing that requires electronic processing at the optical core be
transmitted at a lower data rate (e.g. 2.5Gbps). Figure 3. 1 shows the optical and
electronic functions at different points of an optical network along with issues and
bottlenecks limiting the forwarding process. At the network edge, forwarding is
limited mainly by electronic bottlenecks such as IP packet processing and bit level
frame processing at edge data rates. Furthermore, it is critical that any framing
mechanism used ensures error-free bit level recovery of forwarded payloads. At the
core however, network performance limitations include optical rise/fall time
considerations of forwarding elements, the electronic header recovery and
uncertainties in control signals. Complete control over serialization and
deserialization processes is required to reduce control signal timing uncertainties to
within a bit of the serialized data rate. However, commercially available electronics
30
Ingress
CoreEgress
OPTICS
ISSUESAND BOTTLENECKS
ELECTRONICS
* Fast wavelength tuning
* Fast wavelength tuning* Packet Wavelength Conversion
* Optical Rx
* Parallel processing of serial data
* Optical rise/fall times* Electronic header CDR* DC balance
* Per packet payload CDR* Physical layer variations
* Framing* Serialization
* Header recovery* Forwarding control signal generation
* Deframing* Deserialization
OPTICAL NETWORKIngress
CoreEgress
OPTICS
ISSUESAND BOTTLENECKS
ELECTRONICS
* Fast wavelength tuning
* Fast wavelength tuning* Packet Wavelength Conversion
* Optical Rx
* Parallel processing of serial data
* Optical rise/fall times* Electronic header CDR* DC balance
* Per packet payload CDR* Physical layer variations
* Framing* Serialization
* Header recovery* Forwarding control signal generation
* Deframing* Deserialization
OPTICAL NETWORK
Figure 3. 1: End-to-End network functional architecture.
do not satisfy this new requirement as they are traditionally used on continuous data
streams.
3.1.1 Traffic characteristics at the Network Edge
Packet traffic entering and leaving an OPS network is framed in existing data
link/physical layer protocols such as Ethernet or Packet over SONET (PoS).
Connection between the optical network edges may be viewed as simple IP only
connections with control of packet forwarding within the optical network confined to
the optical control plane as in the case of a peer-to-peer network. Alternately, it may
be dictated by the routing mechanism external to the optical network as would be the
case of an overlay model (Figure 3. 2). In order for the optical network to be protocol
31
transparent at the network edge i.e. to allow for IP packet traffic to enter/leave the
OPS network
Optical control plane and forwarding plane
External Network Control Plane
Optical control plane and forwarding plane
External Network Control Plane
(a)
External Network Control Plane
External Network Control Plane
(b)
Figure 3. 2: a) Peer-to-Peer vs. (b) Overlay optical network approach
over various lower layer protocols, any protocol connection into or out of the OPS
network must be terminated at the edge. In this thesis, we focus primarily on IP
packets transported over PoS interfaces as the connection to the outside network. A
suitable framing/deframing card at the ingress/egress of the OPS network would
allow for interfacing to any other external protocol standard. IP Packets are assumed
32
to be of variable byte size from 40 to 1500 bytes and fragmentation of packets is not
considered due to the additional complexity and problems associated.
3.1.2 Electronics in optical networks
While traffic within an OPS network predominantly remains in the optical domain,
control of all traffic is performed using electronics. At the network edges, processing
and framing of the entire packet is done in parallel words of 32/64 bits or higher.
However, unlike most today’s transport standards where processing is performed on
the entire stream of packets, each transmitted packet may follow a different
forwarding path through the network. Consequently, bit level processing on parallel
word traffic would have to be performed considering each packet individually as
opposed to on the entire data stream for traffic on existing standards. Furthermore,
aggregation of packets may be required to scale from lower edge rates for better
utilization of high-speed optical links. During detection at the network egress, per
packet variation of the optimum bit threshold level and the appropriate choice of
sampling instant is critical to maintain error free operation when receiving packets
from different sources. Limiting amplifiers may be used to provide suitable
interfacing between detected analog signals and the digital sampling circuits. At the
network core, a weak temporal relationship between the payload and framing exists.
Electronic rise/fall times of control signals add up to rise/fall times of optical
components while timing uncertainty caused due to serialization/deserialization
originates primarily from core node electronics. These factors must be taken into
account while designing the optical framing mechanism. Control electronics also
33
convert digital control bit into analog signals for control of optics and include Digital-
to-Analog Converters (DAC), limiting amplifiers, modulator drivers and current
drivers. After header removal, control signals are generated at packet rate. Varying
interpacket gaps result in varying dc content in these control signals and pose a
problem in electronics as most commercial high speed electronics are ac coupled. A
compromise between rise/fall times and ac coupling effects may be required for
proper operation. Advantages in packet level processing of forwarding information at
the optical core nodes may be counterbalanced by an inefficient framing mechanism.
Guard bands and synchronization times that allow for processing time in the optical
domain decide the efficiency of the packet framing. However, at the cost of lower
link bandwidth utilization, decoupling of high bit rate payloads from the base rate
framing using expensive guard bands can result in significantly greater scalable
packet forwarding rates.
3.2 Optical Framing Considerations
In this section we design and implement a framing structure based on various
requirements by both the optical core and edge electronic processing stages to ensure
a) proper all-optical payload forwarding and b) successful recovery of all the payload
bits. The general packet structure consists of a framing mechanism containing header
and packet delineation words encapsulating a decoupled payload that may be at
higher bit rate. Decoupling ensures payload transparency in that payloads may be
processed all-optically at the core after header and optical framing information have
34
been stripped and based only on optical header information. The framing mechanism
must support continuously variable payload sizes from 40 to 1500 bytes. After
framing, measurements within the optical network would be confined to the physical
layer whereas layer-3 measurements may be performed end-to-end only.
3.2.1 Forwarding requirements
Edge forwarding may be based on the raw IP header while at the core examination of
the optical payload is not possible. Forwarding information must therefore be
included in the optical header and may be extracted from the IP destination address,
the Type of Service (TOS) and the Time To Live (TTL) fields. It is desirable that
although the IP destination address is 32 bits long, the size of the Optical Header
(OH) be kept small for fast next hop and label lookup at the core. While this limits the
effective maximum size of the optical network, label reusability in label swapping
technologies helps in extending the network size further. Quality of service and
differentiated services may be implemented by including a QOS field in the optical
header. While this information may be based on the TOS field of IP packets, tighter
network control is possible as these bits are set by the optical ingress nodes and not
by the end user as is the case in IP. This could affect bandwidth provisioning and core
node decisions such as access to buffering and contention resolution elements. The
TTL field in an optical packet may be either transported end-to-end and decremented
at the optical egress point, or may be decremented at each node within the optical
network and used to ensure that packets do not occupy network bandwidth infinitely
due to virtual forwarding loops. When the TTL field falls to zero and the packet does
35
not reach a network egress, packet payload may either be re-examined and a new
optical header assigned, or the packet may be simply dropped.
3.2.2 Idlers
In OPS, each optical packet may be viewed as an individual routable unit and packets
in each stream may be statistically multiplexed and demultiplexed at each node within
the optical network. Optical receivers both at the core and egress nodes however,
require that the average power level on incoming link be kept constant for proper
operation. DC balance of the optical packet becomes critical for the following
reasons:
1. Equalization of optical power through the entire packet. Lack of power
equalization would lead to uneven gain at EDFA (amplifiers).
2. Assurance of a bit transition required for clock recovery every specified number
of bits. For any clock recovery system to work, a guaranteed bit transition is
required over a certain number of bits, to charge the system. DC balance can
provide this guarantee.
Idlers or inter-packet fills may be added between packets in a stream to maintain
constant average power and provide adequate dc balance between packets. However,
a proper encoding scheme is required to provide DC balance within the optical packet
itself, more specifically within the optical payload. Standard encoding formats such
as 64B/66B or Manchester encoding can be used to provide the required DC balance.
Idlers may also be used for link level communication between nodes. Characters used
as idlers must be uniquely identifiable and must contain sufficient bit transitions to
36
maintain dc balance. At the optical ingress, idlers may be generated locally and
inserted after payload processing and header insertion as the entire packet stream is
transported together. At the network core however, statistical multiplexing and
demultiplexing of packets changes the nature of packet streams and external idlers
may be required to fill newly generated gaps within the packet stream. In the absence
of proper dc balance in the core, optical components such as limiting amplified
photoreceivers and amplifiers experience transient behavior as shown in Figure 3. 3
that may corrupt the packet stream and result in permanent loss of information.
Amplifier Gain Transients
Photodetector/TLADetector/TLA
slow start
AmplifierIncoming packet stream
Packet distortion
Amplifier Gain Transients
Photodetector/TLADetector/TLA
slow start
AmplifierIncoming packet stream
Packet distortion
Figure 3. 3: Need for idlers - transmission distortions in the absence of dc balance
3.2.3 Edge and Core node Clock-Data Recovery (CDR)
After photodetection, successful recovery of bits within the optical packets depends
on clock and data recovery. After forwarding, packets arriving at the next hop may be
from different node transmitters and consequently at slightly different phase and
frequencies. Moreover, since optical framing is erased and rewritten at each hop,
there is no correlation between phase and frequency of frame and payload clocks.
Clock recovery must therefore be performed on the fly and with extremely short
capture times. An OPS framing mechanism must therefore allow for packet level
37
clock synchronization times and must ensure adequate bit transitions within both the
payload and the framing to facilitate clock tone isolation from the data stream. At the
core, clock synchronization achieved at the start of the optical packet is lost during
the payload time and therefore must be locked on again before the end of optical
packet is detected. While all optical clock recovery approaches have been proposed
[1], a filter based approach is proposed for clock recovery as unlike Phase Locked
Loops (PLLs) and flip-flop based bang-bang detectors (usec range capture times),
this approach can capture clock information within tens of bits. Figure 3. 4 shows the
general structure of a filter based clock recovery approach.
Photodetect
Amplifier
Frequency doubler
Narrowband Filter
Limiting amplifier
Photodetect
Amplifier
Frequency doubler
Narrowband Filter
Limiting amplifier
Figure 3. 4: Packet mode clock recovery
In the case of Non-Return-to-Zero (NRZ) data transmission as is usually the case with
lower bit rates used for optical framing, a frequency doubler is used to retrieve the
appropriate clock tone, whereas this may be excluded when CDR is performed on
Return-to-Zero (RZ) traffic. After initial photodetection and amplification of the
incoming data signal, a doubler is used to recover the clock tone at the data repetition
rate. A narrowband filter is then used to reject all frequencies but the clock tone. The
Q value of the filter decides the capture time of the clock recovery circuit and the
higher the Q, the faster the capture of clock information. However a tradeoff in Q
38
value selection is preferred, as a faster capture time also means a smaller hold time
after data bit transitions are removed. The recovered clock signal suffers from
amplitude variations due to the varying clock content within the data bits and
therefore necessitates the use of a limiting amplifier with a wide input amplitude
dynamic range to guarantee a stable clock. At the core, any optical adaptation to
different bit rates would require additional payload clock recovery. However, this
may be limited to a sub harmonic clock recovery by use of appropriate synch words
and encoding schemes.
3.2.4 Framing structure
The AOLS adaptation layer must satisfy the following main criteria:
1. Complete optical payload and frame decoupling must be achieved. Packet
level processing of optical payload based only on the optical packet header
must be possible and proper delineation of both packet and payload is
required.
2. As discussed in the previous section, on the fly, clock recovery must be
possible based on synchronization words both at the header and the payload.
Adequate frame and payload coding is required to ensure sufficient bit
transitions.
3. All forwarding information for optical packets must be self contained within
the header and must be accessible lower bit rates for electronic processing at
the optical core nodes. All forwarding decisions made at the core are based
39
completely on this information without requiring electronic examination of
the optical payload.
4. Interpacket fills must be in place at all times to ensure adequate dc balance.
Idlers used for this purpose must contain sufficient bit transitions to maintain a
constant average power.
5. Any further optical adaptation requiring transfer of information on the payload
must be contained within the optical header and passed along with the payload
even after label swapping.
Based on these requirements we propose an AOLS packet framing mechanism as
shown in Figure 3. 5.
IdlersIdlers SOOP Optical PayloadSW OL SW SOIP EOIP SW EOOPGB GB IdlersIdlers SOOP Optical PayloadSW OL SW SOIP EOIP SW EOOPGB GB
Variable sized decoupled payload container
Optical Packet
Optical packet framing Optical packet framing
IdlersIdlers SOOP Optical PayloadSW OL SW SOIP EOIP SW EOOPGB GB IdlersIdlers SOOP Optical PayloadSW OL SW SOIP EOIP SW EOOPGB GB IdlersIdlers SOOP Optical PayloadSW OL SW SOIP EOIP SW EOOPGB GB
Variable sized decoupled payload container
Optical Packet
Optical packet framing Optical packet framing
Figure 3. 5: Optical Framing Structure. SW - Synch. Word; S/EOOP –Start/End Of Optical Packet; OL - Optical Label; S/EOIP- Start/End Of IP Packet
A synchronization word (SW) at frame data rate is used at the start of each optical
packet for clock recovery followed by a unique Start Of Optical Packet (SOOP)
identifier word. An End Of Optical Packet (EOOP) identifier demarcates the end of
each optical packet and is preceded by another frame rate synchronization word. All
forwarding information required within the optical network is contained in the
Optical Label (OL). Guard bands are used to temporally decouple the optical payload
from the framing and the size of the guard bands is determined by several factors
40
including the largest rise/fall time for optical processing elements within the core and
any accumulative effects of timing uncertainties. Each packet payload container starts
with a payload rate synchronization word and a Start Of IP Packet (SOIP) identifier
word and may be of variable length. An End Of IP Packet (EOIP) word denotes the
end of the optical payload and no synchronization word is required preceding this as
clock information is assumed to be maintained throughout the payload. Each optical
payload undergoes a simplified 32B/34B encoding scheme to ensure that unique
framing identifiers are not present within the payload. Each word consists of 32 bits
or 4 bytes to conform to the electronic processing bus width and each optical packet
has a 10 words or 40 bytes overhead. Figure 3. 6 shows the simple structure of the 32
bit optical header which consists of 12 reserve bits that may be used to carry
additional information for any optical adaptation, 9 bits of an Optical identifier
(O_ID) used in making forwarding decisions at the core, 8 bits for TTL and 3 bits for
priority to implement quality of service.
O_ID (9 bits)
TTLOptical(8 bits)
Priority (3 bits)
Reserve bits (12 bits)
Optical Label – 32 bits
O_ID (9 bits)
TTLOptical(8 bits)
Priority (3 bits)
Reserve bits (12 bits)
Optical Label – 32 bits
Figure 3. 6: Example 32 bit Optical Label or header. O_ID - Optical Identifier; TTL - Time To Live
While 9 O-ID bits limit the number of unique ID combinations to 512, label reuse
functionality of the label swapping technique can be used to increase the effective
maximum optical network size significantly. Figure 3. 7 shows scope traces of the
framed optical packets transmitted from the optical ingress.
41
Optical PacketsOptical Packets
(a)
SOOP EOOP
SOIP EOIP
GBSOOP EOOP
SOIP EOIP
GB
(b)
Figure 3. 7: Experimental scope traces of (a) Optically framed packets (b) Packet framing (refer Figure 3.6)
3.3 Edge Node Adaptation for OPS Traffic
While prior state of art framing mechanisms deal with fixed sized packets [2], the
AOLS adaptation considers forwarding requirements of variable sized packets. Edge
node adaptation acts as the interface between electronic routing (with optical
transport such as 10GigE and PoS) and all-optical packet forwarding. This section
deals with the structure of edge nodes to successfully implement the adaptation
mechanism proposed in the previous section. Figure 3. 8 shows the adaptation layer
processing steps at the optical network edges. Raw IP packets are extracted from
external transport protocol framing and undergo encoding and delimiting to form the
optical payload. Forwarding information extracted from the IP header is then used to
generate an optical header and framing before the optical packet is placed in the
stream. Locally generated idlers are finally inserted into the data stream to ensure
uniform average power before transmission into the optical network at the appropriate
42
wavelength. After all-optical forwarding through the optical network, payload
extraction from the optical data stream is performed before
IP Packet
Encoded IP
Encoding
Payload Framing
Optical Packet Framing
Idler Insertion
Framed Payload
Framed Optical Packet
Idlers Inserted data stream
Optical λTuning
Payload Extraction
Packet id and Idlers Extraction
Optical Receive
Encoded IP extraction
Decoding
All-Optical Packet Switching
INGRESS EGRESS
Physical layer Transmission
IP Packet
Encoded IP
Encoding
Payload Framing
Optical Packet Framing
Idler Insertion
Framed Payload
Framed Optical Packet
Idlers Inserted data stream
Optical λTuning
Payload Extraction
Packet id and Idlers Extraction
Optical Receive
Encoded IP extraction
Decoding
All-Optical Packet Switching
INGRESS EGRESS
Physical layer Transmission
Figure 3. 8: Edge node adaptation layering
payload decoding and IP packet extraction takes place. The IP header may be updated
with information such as TTL from the optical packet header before the IP packets
are framed in external transport protocol framing and exit the optical network. It is
important to note that while data link and physical layer link signaling are terminated
at the optical network edge, layer-3 or IP continuity is maintained through the AOLS
network and is required for end-to-end performance measurements.
3.3.1 Ingress Node Description
The ingress node functionality implements the following set of steps to generate an
optical packet stream.
43
1. Deframe incoming data from external protocols and present raw IP packets
ready for packet adaptation.
2. Process IP packets and generate a framing by encapsulating each packet into
an optical packet. This stage also performs label lookup based on the IP
header and payload encoding.
3. Fast tunable laser with digital control input to transmit each optical packet at
the appropriate wavelength and an optical modulator that generates an optical
data stream based on the framed electrical packet stream. Wavelength
selection is required to perform space switching by use of an Arrayed
WaveGuide Router (AWGR).
POS deframe
Layer-3 lookup and packet
frame
Fast tunable laser
Optical modulator
POS traffic in
Optical packet traffic out
INGRESS
POS deframe
Layer-3 lookup and packet
frame
Fast tunable laser
Optical modulator
POS traffic in
Optical packet traffic out
INGRESS
Figure 3. 9: Ingress node high level schematic
Figure 3. 9 shows the functionality of the ingress node. Incoming traffic is received as
a PoS OC-48 (2.488Gbps) stream and is decapsulated in the POS deframer which
outputs raw IP packets at its SONET PHY interface. A General Purpose Protocol
Processor (GPPPi) performs individual IP packet identification lookup and framing
along with locally generated idler insertion. A fast tunable laser board controlled by
44
the GPPPi is used to generate CW of the wavelength selected on a per packet basis
onto which optical packets are modulated using an electro-optic modulator. The
ingress node may be adapted to receive IP packets in any external transport protocol
by simply exchanging the POS deframer board with a suitable decapsulation board.
3.3.2 Egress Node Description
Egress node functionality is implemented in the following steps to extract IP packets
and forward them out of the optical network.
1. Amplified photodetection followed by clock and data recovery for payload
(and framing).
2. Payload identification and IP packet extraction.
3. PoS Framing before IP packets exit the optical network.
After successful forwarding through the optical network, packets arrive as an optical
data stream into the egress node (Figure 3. 10).
POS deframe
Optical packet detect
deframe
Photodetectand CDR POS traffic out
Optical packet traffic in
EGRESS
POS deframe
Optical packet detect
deframe
Photodetectand CDR POS traffic out
Optical packet traffic in
EGRESS
Figure 3. 10: Egress node functionality. CDR- Clock Data Recovery
While optical framing is generated by a single upstream node for all optical packets,
optical payloads maybe have undergone completely different processing stages and
may originate from different ingress nodes. Consequently payloads may be of varying
45
amplitudes and signal qualities. A limiting amplified photodetector is used at the
egress to eliminate amplitude variations between packet payloads and present a
constant electrical stream for Clock Data Recovery (CDR). At the egress, optical
framing may or may not be required to be examined and could therefore require
separate CDRs for the frame and payload data extraction. Clock for data recovery is
fed from the transmitter ingress node to core and egress nodes for all demonstrations
and performance measurements performed although a filter based clock recovery
technique is proposed in this research. At the GPPPe, IP packet extraction is
performed after suitable payload identification based on optical frame demarcation
words and decoding of the payload container. Extracted IP packets are finally
encapsulated into OC-48 PoS traffic using a PoS framer board before exiting the
optical network.
3.4 Edge Node Implementation
Edge node implementation is carried out using three main boards: the POS deframer,
the GPPP and the digitally controllable fast tunable laser board. Electronic to Optical
(E/O) and Optical to Electronic (O/E) functions are performed by use of a Lithium
Niobate (LiNbO3) modulator and a limiting amplified photodetector both with a with
a 10GHz bandwidth. Payload decoupling is achieved by implementing the optical
packet structure discussed in section 3.2.4 and can support forwarding of high bit rate
payloads with lower bit rate optical framing. However, in this research, emphasis is
placed on examination of implementation bottlenecks and performance of AOLS in
46
forwarding IP traffic. Payload and framing data rates are therefore chosen to be the
same at 2.5Gbps although core processing examines only optical framing information
for forwarding decisions.
3.4.1 PoS Framer/ Deframer
A Vitesse POS (Packet Over SONET) physical layer framer/deframer board operating
at 2.488Gbps was used to perform SONET and PPP/HDLC framing/deframing of the
IP packets traveling in and out of the AOLS network. The board consists of the POS
line optical interface, a VSC8140 clock-data recovery unit and a VSC 9112 POS
framer/deframer chip as shown in Figure 3. 11.
Photodetect VSC 9112 Framer/deframer
Stratum -3Crystal
oscillator
VSC 8140 CDR
16
16
32
32
Tx
Rx
Raw IP Packet
Interface (PIF)
Photodetect VSC 9112 Framer/deframer
Stratum -3Crystal
oscillator
VSC 8140 CDR
16
16
32
32
Tx
Rx
Raw IP Packet
Interface (PIF)
Figure 3. 11: Vitesse POS OC-48 Framer/Deframer functionality
Traffic entering the AOLS network is detected by the onboard photoreceiver and
clock data recovery is performed by the VSC8140 unit using a locally generated
stratum-3 77.78MHz clock from a crystal oscillator. Digital data bits then arrive at the
line side POS interface and are examined by the 9112 deframing section. SONET
47
overhead information is extracted from the POS frames and PPP/HDLC framed IP
packet stream is forwarded to the next section. Individual IP packets are then
extracted from the PPP/HDLC stream at the packet demapping section. The extracted
IP packets are queued in their order of arrival at the line side interface buffer, ready
for encapsulation into the optical packet format by the ingress General Purpose
Protocol Processor (GPPP). The IP packet line side interface may be run by a
77.78MHz clock from the GPPP independent from the POS interface side thereby
providing separation of clock domains between the synchronous SONET world and
the AOLS optical domain. The IP line side buffers are capable of satisfying any
buffering requirements arising due to slight clock frequency misalignments. The
framer/deframer card may be appropriately configured for the type of POS traffic
received/ transmitted via a computer interface. While the size of the current board
used is about 16”x12”, the board may be reduced to less than a fourth of this size for
commercial use.
3.4.2 Electronic Control design
General Purpose Protocol Processors (GPPP) are used to adapt raw IP packets to and
from the optical framing mechanism. IP packets from the POS framer/deframer board
are transferred to and from the GPPP via a 32 bit wide parallel bus clocked at
77.78MHz. Figure 3. 12 shows the ingress FPGA process functionality where
incoming IP packets are transported along with control signals. IP header processing
and lookup is performed on a per packet basis to determine the appropriate optical
header and wavelength select. Optical payloads then undergo encoding to eliminate
48
the accidental presence of a control word such as SOOP, EOOP and are finally
encapsulated into an optical frame before idler insertion into the stream.
IDENTIFYIP HEADER
INFORMATION
LOOKUP LABEL &
λ
FRAMEIP
PACKET ENCODER IDLERFILL
RAW IP (32)FROM POSDEFRAMER
CONTROL (5)
PROCESSEDDATA (32)
CONTROL (4)
λ Ingress (8)
ENCODEDDATA (32)
CONTROL (1)
ENCODED DATA-STREAM WITH INTER-PACKET
IDLERS (32)
INGRESS FPGA
OPTICAL LABEL (32)
PROCESS STAGE
IDENTIFYIP HEADER
INFORMATION
LOOKUP LABEL &
λ
FRAMEIP
PACKET ENCODER IDLERFILL
RAW IP (32)FROM POSDEFRAMER
CONTROL (5)
PROCESSEDDATA (32)
CONTROL (4)
λ Ingress (8)
ENCODEDDATA (32)
CONTROL (1)
ENCODED DATA-STREAM WITH INTER-PACKET
IDLERS (32)
INGRESS FPGA
OPTICAL LABEL (32)
PROCESS STAGE
Figure 3. 12: Ingress FPGA process functional diagram
GPPPi generates a 2.5Gbps electrical optical packet stream and a set of parallel
‘wavelength select’ bits to control the fast switching tunable laser board. While
encoding used is a simple principle of embedding 0’s in a word of 32 bits within the
payload, the implementation is complex due to the following reasons:
1. 34 bits are generated for every 32 bits input therefore requiring a degree of
flow control to fit encoded payload bits into a 32 bit wide bus.
2. Bit level operations on a 32 bit word based stream are necessary. Rotation of
the entire packet payload stream on a word by word basis is used to emulate
bit stream traffic and achieve encoding.
3. Since encoding is done only on packet payloads, it must start new for each
incoming optical payload. Precise control signaling is therefore necessary to
achieve this as opposed to encoding on a continuous bit stream as is the case
with existing protocols.
49
Egress node electronic control performs the reverse process of identification of
optical payload, decoding and extraction of raw IP packets. Decoupling of payload
containers from the optical packet framing results in payloads arriving at random bit
position relative to the 32 bit processor boundary. Figure 3. 13 shows the egress
FPGA process functionality which includes a phase locker to align payload containers
to the 32 bit processing word boundary, a decoder, and a IP packet extraction stage to
assemble the entire IP packet before transferring it to the POS framer.
Raw DMUX signal bus
(32)
PHASE LOCKER
PAYLOAD CONTAINER
DETECTDECODER
CLOCK (1)
PROCESSEDPAYLOAD (32)
CONTROL (7)
RAW IP Packets(32)
EGRESS FPGAPOS PHY control
(5)
PHASE LOCKEDENCODED DATA (32)
Raw DMUX signal bus
(32)
PHASE LOCKER
PAYLOAD CONTAINER
DETECTDECODER
CLOCK (1)
PROCESSEDPAYLOAD (32)
CONTROL (7)
RAW IP Packets(32)
EGRESS FPGAPOS PHY control
(5)
PHASE LOCKEDENCODED DATA (32)
Figure 3. 13: Egress FPGA process functionality
The phase locker compares incoming sequential 32 bit words with a SOIP and EOIP
words to identify one of 31 bit phases in which the incoming payload may be present
as shown in Figure 3. 14. An aligner stage uses phase information to rotate the
incoming payload to the 32 bit word boundary for further processing. Decoder
performs the reverse process of the encoder and results in 32 valid payload bits for
every 34 bits received. IP packets must therefore be stored in a FIFO until the entire
packet has been decoded before they are ready for transfer into the POS framer.
50
32 to 64 BIT CONVERTER
RAW INPUTDATASTREAM
(32)
64 BITRAW INPUT
DATASTREAM(64)
…
4 3 2 1
SOIP WORD NOT ALIGNED TO 32 BIT BOUNDARY
…1
2
2
3 ...
SOIP/EOIP DETECTOR
PHASE 1
PHASE 2
PHASE 31
ALIGNER/ OPTICAL DECAP-
SULATION
64 BITRAW INPUT
DATASTREAM(64)
…
4 3 2 1
SOIP WORD ALIGNED TO 32 BIT BOUNDARY
PHASE ALIGNMENTCONTROL SIGNALS (36)
CONTROL SIGNAL(5)
ALIGNED OPTICALPAYLOAD (32)
32 to 64 BIT CONVERTER
RAW INPUTDATASTREAM
(32)
64 BITRAW INPUT
DATASTREAM(64)
…
4 3 2 1
SOIP WORD NOT ALIGNED TO 32 BIT BOUNDARY
…1
2
2
3 ...
SOIP/EOIP DETECTOR
PHASE 1
PHASE 2
PHASE 31
ALIGNER/ OPTICAL DECAP-
SULATION
64 BITRAW INPUT
DATASTREAM(64)
…
4 3 2 1
SOIP WORD ALIGNED TO 32 BIT BOUNDARY
PHASE ALIGNMENTCONTROL SIGNALS (36)
CONTROL SIGNAL(5)
ALIGNED OPTICALPAYLOAD (32)
Figure 3. 14: Egress FPGA - Phase locker operating principle
3.4.3 Tunable Laser board and control
A fast tunable laser board was specially designed to control a pigtailed 3-section
Bookham Distributed Feedback (DFB) laser to tune between selected wavelengths
within a few nanoseconds. Whereas, commercial laser boards tune in the usec range,
tunability in a few nanoseconds is important to forward packets rapidly and reduce
guard bands between packets. Figure 3. 15 shows the functional schematic of the fast
tunable laser board designed and is digitally controlled using a parallel bus clocked
from the upstream control device. An FPGA is used to convert the digital wavelength
select bits into appropriate DAC values that set the current for the gain, phase and
rear sections. Fast settling DACs from Analog Devices (AD) were used along with an
op-amp based current driver circuitry to tune to the selected current values.
51
FPGA digital current value
lookup
DAC DAC DAC
V/I V/I V/I
DFB Laser
V1 V2 V3
I3I2I1
12 12 12
Digital control
CW out
Digitally controlled fast tunable Laser board
FPGA digital current value
lookup
DAC DAC DAC
V/I V/I V/I
DFB Laser
V1 V2 V3
I3I2I1
12 12 12
Digital control
CW out
Digitally controlled fast tunable Laser board
Figure 3. 15: Fast tunable laser functionality. DAC – Digital to Analog Converter; V/I – Voltage to Current converter
Figure 3. 16 shows a sample wavelength spectrum output of the tuned DFB laser and
a sample switching curve. A 0.6nm filter was used to test and measure switching
times that was measured to be ~6nsecs between channels 1548.51nm and 1549.32nm.
Figure 3. 16: Fast tunable laser spectrum at 1549.32nm and switching scope trace time
52
Switching time of the DFB laser is currently limited by the settling time of the DACs
and may be brought down to a few hundred picoseconds in future.
3.5 Edge Adaptation testing and performance
We used Layer-3 performance metrics to evaluate the performance of the back-to-
back edge node adaptation. Testing was done using a Spirent (Smartbits) SMB6000
chassis with a POS OC-48 tester interface and controlled using SmartWindows
application. While all data may be generated and received by the Layewr-3 tester, a
Cisco GSR 12000 electronic router was used in the receive path to a) test
performance limits of electrical routers as compared with the optical adaptation
implemented and b) as incremental testing since the electronic routers act as
aggregation routers for traffic after optical packet forwarding at the core to multiple
egress node. The signal path used in adaptation testing is shown in Figure 3. 17.
Smartbits SMB 6000 tester
OC-48 POS deframer
Ingress adaptation
OC-48 POS framer
Egress adaptation
Tx
Rx
Optical packet stream after adaptation
Electronic Router
Smartbits SMB 6000 tester
OC-48 POS deframer
Ingress adaptation
OC-48 POS framer
Egress adaptation
Tx
Rx
Optical packet stream after adaptation
Electronic Router
Figure 3. 17: Test path and traffic flow setup
No significant throughput or latency changes were observed when electrical and
optical back-to-back measurements were performed indicating that basic optical
53
transmission had no impact on the measurements. Packet sizes were varied from 40 to
1500 bytes and packet destination IP addresses were configured to match an entry in
the ingress header lookup table.
3.5.1 Impact of optical framing on utilization
A theoretical study of the impact on utilization due to adaptation overheads with IP
payloads varying from 40 to 1500 bytes is shown in Figure 3. 18. It can be seen that
for small packets (40bytes) a 50% penalty is incurred due to the 40 byte overhead
while for large packets (1500 bytes), adaptation penalty is below 5%.
0 200 400 600 800 1000 1200 1400 1600
50
60
70
80
90
100
Theoretical Throughput INGRESS --> EGRESS
Thro
ughp
ut (%
)
Packet Size (bytes)
OptFraming Encoded EncodedPOS
Figure 3. 18: Edge adaptation theoretical throughput calculation
An additional average penalty of ~4% is seen due to the encoding scheme used and is
nearly offset when PoS traffic overheads are taken into account to yield a maximum
utilization of ~95%. This is lower than the 97.5% or 99.5% overheads seen in
54
10GigE or PoS but may be viewed as the price paid for finer granularity of
forwarding control and lower stress on the electronic forwarding plane.
3.5.2 Measured throughput performance
Packet throughput measurements through the edge adaptation were measured for
varying IP packet sizes and compared with theoretical throughput calculations
discussed in section 3.2.4. Throughput was measured to be the maximum utilization
at which packet loss rises above 0.1% of transmitted packets.. Figure 3. 19 shows the
throughput measurements done along with theoretical throughput calculations.
0 200 400 600 800 1000 1200 1400 16000
20
40
60
80
100
Thro
ughp
ut (%
)
Packet Size (bytes)
Ingress Egress Theoretical Ingress-Egress-Electronic Router Electronic Router
Figure 3. 19: Edge adaptation throughput measurement and theoretical comparison
To ensure sufficient statistics, 5 million packets were transmitted for each run before
packet loss evaluation for a given utilization was performed. A binary incremental
testing procedure was used was used to identify the throughput utilization for each
packet size. It is important to note that before any layer 3 measurements were
55
performed, physical layer tested was made to ensure a low Bit Error Rate (BER) was
done through the electronic router and through the edge adaptation with the electronic
router to compare and identify possible throughput bottlenecks. For packet sizes
greater than 128 bytes, no significant difference was observed between theoretical
and measured throughput whereas a ~20% difference was observed for smaller packet
sizes. This reduction in maximum utilization was traced to the electronic router
forwarding process and not due to the adaptation. Throughput performance through
adaptation was therefore verified to follow the theoretical predictions closely.
3.5.3 Measured latency performance
Latency measurements were performed for different packet sizes through the
adaptation, through the electronic router and through the entire traffic path to identify
contributions to the end-to-end latency as shown in Figure 3. 20. It can be seen that
electronic router contributes a significant portion of the measured latency. A 20%
utilization was used to ensure no packet loss during measurement. Latency increase
with packet size is observed to be fairly linear due to the absence of any complicated
loop up mechanism at the ingress and a maximum of ~41usec latency is observed
through the entire traffic path.
56
0 200 400 600 800 1000 1200 1400 1600
5
10
15
20
25
30
35
40
45
50
Late
ncy
(use
c)
Packet Size (bytes)
Electronic Router Ingress-Egress Ingress-Egress-Electronic Router
Figure 3. 20: Edge adaptation latency measurement
3.6 Chapter Summary
In this chapter, framing considerations were examined and a packet framing structure
that accounts for various signal bandwidths and processing time limitations was
proposed and implemented. An optical edge node adaptation capable of supporting bit
rate transparent all-optical payload forwarding based only on framing information
was studied and demonstrated. Implementation sub-systems necessary to build AOLS
Ingress and Egress nodes were examined. The need for payload encoding, per packet
clock recovery, a decoupled payload framing, a self contained optical packet header
and idlers to maintain constant average power were identified and incorporated in the
proposed framing mechanism. We finally look at the end-to-end performance of the
implemented adaptation using layer-3 test metrics such as throughput and latency and
57
find a close match with theoretical predictions in the behavior for almost all packet
sizes. For small packets, throughput performance dropped to ~40% while for larger IP
packets, throughput was measured to be ~95% comparable to 10GigE and PoS
measurements. No significant latency penalty was observed for all packet sizes and
increase in latency with packet size was noted to be linear as expected due to the
simple nature of the header lookup. Electronic processing at line rate was
demonstrated to perform IP header look up, payload encoding, and idler insertion. We
present the first comprehensive implementation of OPS adaptation for variable length
IP packets to date and introduce layer-3 measurement metrics to evaluate the network
performance of such an adaptation mechanism.
58
References
[1] B. Sartorius, "All-optical clock recovery for 3R optical regeneration,"
presented at Optical Fiber Communication Conference and Exhibit, 2001.
OFC 2001 MG7-1-MG7-3 vol.1, 2001.
[2] P. Gambini, M. Renaud, C. Guillemot, F. Callegati, I. Andonovic, B. Bostica,
D. Chiaroni, G. Corazza, S. L. Danielsen, P. Gravey, P. B. Hansen, M. Henry,
C. Janz, A. Kloch, R. Krahenbuhl, C. Raffaelli, M. Schilling, A. Talneau, and
L. Zucchelli, "Transparent optical packet switching: network architecture and
demonstrators in the KEOPS project," Selected Areas in Communications,
IEEE Journal on, vol. 16, pp. 1245-1259, 1998.
59
Chapter 4
All-Optical Label Swapping
Network Performance
A major performance benchmark in forwarding technologies is the number of packets
forwarded through a switch fabric at any given interval of time. All-optical
forwarding promises the advantage of increasing the packets forwarded per second by
relieving the electronic domain of burdensome processing and by reducing
forwarding times using high bit rate optical payloads. Moreover, the packet level
granularity in packet switching offers the possibility of a high degree of resource
sharing and consequently a better performance in bursty internet traffic seen today.
Electronic framing and packet adaptation necessary for enabling OPS was
investigated in the previous chapter. Forwarding variable sized IP packets using
existing optical technologies can now be investigated using traffic from and to the
optical edge nodes in place. In this chapter, we take a look at enabling optical
technologies such as wavelength conversion and the associated control electronics
necessary to implement asynchronous packet forwarding at the optical core. While
physical layer measurements may be made to characterize the various subsystems
within the optical core, layer-3 metrics are critical in evaluating the true performance
60
of both the optical forwarding mechanism within the core and the effectiveness of the
adaptation layer implemented at the network edge. Finally, issues that limit the
network performance are identified and suggestions to improve overall network
throughput are considered.
4.1 All-Optical forwarding using Wavelength Conversion
The switch fabric is the central point of any router and sets a cap on the best
performance achievable. In today’s electronic routers, switching may either be
performed by transporting the entire packet between input and output queues
(implemented using RAM buffers) (e.g. Cisco GSR) or by simply switching tags
containing the address of the location of packets (Juniper M40). In both cases
however, entire packets must be moved between buffers to perform forwarding, an
operation that consumes considerable energy. All-Optical packet forwarding presents
the unique concept of forwarding payloads without expending much energy to store,
examine or switch between input and output ports by decoupling forwarding
information from the entire payload. One way of implementing this is by space
switching optical payloads based on their wavelengths. An Arrayed WaveGuide
Router (AWGR) is a passive optical component that achieves space switching by
exploiting the additional wavelength dimension in optics to as shown in Figure 4. 1.
An AWGR maps input port, wavelength pairs to specific output port, wavelength
pairs in a cyclic fashion and may be viewed as a non-reconfigurable, non-blocking
61
switch fabric. While dynamic switching of optical packets may not be possible using
an AWGR alone, a fast tunable wavelength converter at the input ports of the AWGR
λ11,λ2
1, λ31, λ4
1,
λ12,λ2
2, λ32, λ4
2,
λ13,λ2
3, λ33, λ4
3,
λ14,λ2
4, λ34, λ4
4,
λ11,λ2
4, λ33, λ4
2,
λ12,λ2
1, λ34, λ4
3,
λ13,λ2
2, λ31, λ4
4,
λ14,λ2
3, λ32, λ4
1,
4 x 4 AWGR
1
2
3
4
1
2
3
4
λ11,λ2
1, λ31, λ4
1,
λ12,λ2
2, λ32, λ4
2,
λ13,λ2
3, λ33, λ4
3,
λ14,λ2
4, λ34, λ4
4,
λ11,λ2
4, λ33, λ4
2,
λ12,λ2
1, λ34, λ4
3,
λ13,λ2
2, λ31, λ4
4,
λ14,λ2
3, λ32, λ4
1,
4 x 4 AWGR
1
2
3
4
1
2
3
4
Figure 4. 1: 4 x 4 AWGR wavelength mapping - passive, non-blocking, static optical switch; λji – input port i, wavelength j
can be used to implement dynamic space switching. Packets entering an input port
may be steered to any physical output port by simply wavelength converting them to
the appropriate wavelength before propagation through the AWGR. Packets can be
converted to outgoing wavelengths at the output ports prior to exiting the node. Fast
tunable wavelength converters (TWC) in compliment to AWGRs may therefore be
viewed as the optical equivalent of an electronic switch fabric. However, unlike
electronic switch fabrics implemented using RAM buffers, efficient packet queuing is
not possible in the optical switch fabric due to the absence of optical random access
buffer. Fixed fiber delay lines may however be used as buffering technologies along
with the optical switch fabric to reduce packet collisions during switching and
improve performance.
62
4.1.1 Statistical Demultiplexing
IP packets arriving at the network ingress of an AOLS network undergo edge node
adaptation and are encapsulated in an optical framing mechanism before being
injected into the optical network. At each node within the optical network, packets in
the stream are individually forwarded to output ports based solely upon forwarding
information contained in the optical packet header. When two or more packets
arriving at the core node switch request to be forwarded to the same output port that
cannot be shared, packet contention occurs.
Input
Outputs
Input
Outputs
(a)
InputOutputs
InputOutputs
(b)
Figure 4. 2: Statistical demultiplexing (a) 1 input (b) 4 inputs - forwarding without contention resolution
63
We consider the simplest case of a single input packet stream at the core distributed
to multiple output ports based on packet forwarding requirements and a core node
with no contention resolution mechanism may be viewed as independent input packet
streams being statistically demultiplexed to multiple output ports as in Figure 4. 2.
4.1.2 Core node issues in optics and electronics
Forwarding rate at the optical core is influenced by several electronic and optical
factors. While a TWC and AWGR provide a switch fabric for the optical core node,
switching time of the wavelength converter laser is critical in determining the
minimum guard band required between incoming packets and consequently the
forwarding rate. Rise/fall times and the granularity of timing control of the electronic
control signals to the switching section of the TWC also contribute to the minimum
guard band requirement. Control signals used within the core, transition at the packet
envelope rate instead of the packet bit rate. Varying stream utilization at the core
node input would therefore directly translate to changes in the frequency content of
the electrical control signals. The slow (packet rate) signal transition rates coupled
with the fast rise/fall time requirements ( high frequency content) presents a new
challenge in designing control electronics for OPS. AC coupling effects in
commercially available electronic amplifiers and drivers may therefore severely limit
the forwarding performance and the ability of the core node to handle varying input
traffic conditions. As with the edge, maintenance of a constant average power both
before and after forwarding presents an additional challenge at the core.
64
4.2 Optical Core Node
The AOLS core node presents a label swapped OPS architecture where in addition to
optical payloads being forwarded all optically based on label information, optical
packet labels are erased and rewritten at each hop. The general principle of operation
of the optical core node is shown in Figure 4. 3.
Electronic frame detection, control signal
generation
Optical processing – label erase, WC,
label rewrite
External Idler generation
.
.
.
AWGR
Space switching
Input
Output 2
Output N
Output 1
Photodetect
Bulk delay
Control signals
Electronic frame detection, control signal
generation
Optical processing – label erase, WC,
label rewrite
External Idler generation
.
.
.
AWGR
Space switching
Input
Output 2
Output N
Output 1
Photodetect
Bulk delay
Control signals
Figure 4. 3: Optical Core Node - Principle of operation
The incoming packet stream is tapped and photodetected for frame recovery. Framing
information is then deserialized and processed electronically to determine packet
position within the stream and the forwarding information from the optical header. In
the optical domain, all framing is erased from the packet stream and individual
payloads are wavelength converted to the selected wavelengths before being space
switched using an AWGR. At the output, framing is rewritten on the statistically
demultiplexed packet streams along with a new optical label used at the next hop.
Framed statistically demultiplexed packets exit the core node at the desired physical
output ports. 2.5Gbps limiting amplified photodetectors from Infineon were used to
detect packet framing and feed a Vitesse VSC 8132 deserializer chip before
65
processing at the core node FPGA. The core electronic controller generates the
following control signals:
• Core Optical Framing Erase signal – Used at the first stage wavelength
conversion and blanking system to optically erase all optical framing from the
incoming optical packet stream.
• λcore Select and Control signal – Controls a fast tunable laser control board
that tunes the second stage wavelength conversion system to the selected
wavelength on a per-packet basis.
• Optical Frame Rewrite signal – Appropriately timed and used to rewrite new
Optical Label and Framing information on the outgoing optical payload
stream exiting the second stage wavelength conversion system.
• Locally generated Idler Control signal – Control the generation of locally
generated idlers that are used to fill voids in the optical packet data stream at
the outputs of the AWGR.
Figure 4. 4 shows the functional representation of the FPGA process steps and as with
the egress node processing, a phase locker stage is used to align the frame start word
to the process word boundary. Recovered framing delimiters are used to determine
the timing information for frame erase and TWC wavelength conversion control
signals generated by the core electronics. The recovered optical header is used to
decide the exit port of each packet by generating an appropriate wavelength select
signal. A new label is also generated for the next hop, based only on the recovered
optical header.
66
λ cor
eC
ON
TRO
L(8
)
CO
NTR
OL
FOR
LO
CA
LLY
GEN
ERA
TED
IDLE
RS
(2)
Figure 4. 4: Core FPGA Process - 4 control signals. OL – Optical Label; S/EOOP – Start/End Of Optical Packet; DMUX – Demultiplexer.
PRO
CES
SST
AG
E
32 B
IT R
AW
SIG
NA
LFR
OM
DM
UX
77 M
Hz
CLO
CK
(1
)
PHA
SE L
OC
KED
STR
EAM
(32)
CO
NTR
OL
(7)
CO
RE
FPG
A
PHA
SE L
OC
KER
IDLE
RS
EOO
PSO
OP
OL
IDLE
RS
OPT
ICA
L PA
YLO
AD
( N
OT
EXA
MIN
ED IN
CO
RE)
BLA
NK
ING
SIG
NA
L (1
)
NEW
LY G
ENER
ATE
DID
LER
SEO
OP
SOO
PN
EW C
OR
EO
L
FFFF
FFFF
( O
PTIC
AL
PA
YLO
AD P
ASS
THR
OU
GH
)
REW
RIT
E SI
GN
AL
(32)
NEW
LY G
ENER
ATE
DID
LER
S
λ cor
eC
ON
TRO
L(8
)
CO
NTR
OL
FOR
LO
CA
LLY
GEN
ERA
TED
IDLE
RS
(2)
BLA
NK
ING
SIG
NA
L (1
)
BLA
NK
ING
SIG
NA
L (1
)
PRO
CES
SST
AG
E
32 B
IT R
AW
SIG
NA
LFR
OM
DM
UX
77 M
Hz
CLO
CK
(1
)
PHA
SE L
OC
KED
STR
EAM
(32)
CO
NTR
OL
(7)
CO
RE
FPG
A
PHA
SE L
OC
KER
IDLE
RS
EOO
PSO
OP
OL
IDLE
RS
OPT
ICA
L PA
YLO
AD
( N
OT
EXA
MIN
ED IN
CO
RE)
BLA
NK
ING
SIG
NA
L (1
)
NEW
LY G
ENER
ATE
DID
LER
SEO
OP
SOO
PN
EW C
OR
EO
L
FFFF
FFFF
( O
PTIC
AL
PA
YLO
AD P
ASS
THR
OU
GH
)
REW
RIT
E SI
GN
AL
(32)
NEW
LY G
ENER
ATE
DID
LER
S
67
4.2.1 Two-stage wavelength conversion principle
Fast tunable wavelength conversion may be performed in a single stage within the
core node. However, two-stage wavelength conversion offers the following distinct
advantages over single stage conversion:
• A two stage approach can utilize a robust first stage that is polarization
insensitive and is relatively insensitive to the average power level of the
incoming packets. Payloads entering the optical core node may originate from
different ingress nodes and therefore may possess different optical
characteristics.
• Signal regeneration is possible as the second stage of the wavelength
converter may be optimized for best performance for signals from the first
stage and zero power penalty through the wavelength converter may be
achieved [1].
• Cascading two stages makes any-to-any wavelength conversion including
same-to-same wavelength conversion possible.
We use a Semiconductor Optical Amplifier (SOA) based first stage and a Mach
Zehnder Interferometer (MZI) based wavelength converter as the second stage. The
first stage WC uses the Cross Gain Modulation (XGM) process and is therefore fairly
insensitive to input wavelength. While the first stage statically converts incoming
packets from any wavelength to an internal wavelength, the second stage uses Cross
Phase Modulation (XPM) to achieve dynamic wavelength conversion from the
68
internal wavelength to a selected output wavelength on a per packet basis as shown in
Figure 4. 5.
SOA - WC1 MZI – WC2
λinput λoutputλinternal
BPF @λinternal
SOA - WC1 MZI – WC2
λinput λoutputλinternal
BPF @λinternal
Figure 4. 5: 2-Stage wavelength conversion principle. BPF - Band Pass Filter; SOA - Semiconductor Optical Amplifier; MZI - Mach Zehnder Interferometer
4.2.2 First stage conversion and label erase
Another advantage of using an XGM based SOA WC as the first stage is the
possibility of performing label erasure in addition to 1st stage WC. Figure 4. 6. shows
the working of the SOA based first stage WC.
SOA - WC1
λinput λinternal
BPF @λinternal
CW@ λinternal
Packet @ λinput Payload @ λinternal
SOA - WC1
λinput λinternal
BPF @λinternal
CW@ λinternal
Packet @ λinput Payload @ λinternal
Figure 4. 6: SOA based 1st stage Wavelength Conversion and frame erase
69
Incoming packets enter the SOA along with the internal CW light and impose a
modulation on the CW through the gain medium. A notch filter at the internal
wavelength may be used to reject the incoming packet and pass only the wavelength
converted payload at the internal wavelength. Label erasure may be performed by
turning on the SOA gain during the packet payload and turning it off during packet
framing. The XGM process is an inverting process and therefore requires that the
second stage be operated in the inverting regime to obtain the same payload polarity
at the node output. A fast switching current controller board is required to
enable/disable optical frame erase at the first stage wavelength conversion system.
The general schematic of the current controller board is shown in Figure 4. 7.
vin
R1
R2
R3
R4
RLoad (SOA)
R5
R6+- +
-
vin
R1
R2
R3
R4
RLoad (SOA)
R5
R6+- +
-
Figure 4. 7: Op Amp based Fast switching current driver circuitry for SOA.
This board must be capable of handling all bandwidth utilizations and packet length
and therefore need to be dc coupled both at the input and outputs while still supplying
sufficient constant current to switch the SOA. AD8009 high speed (1GHz) current
feedback op amps were used to supply switching current to the SOA stage. The
70
circuit was built to ensure constant current supply through the SOA as its resistance
changes during the turn on cycle. A Kamelian SOA used for the first stage
wavelength conversion and frame erase has the following gain response with input
power. The SOA along with the fast switching current controller board was tested for
response to applied input electrical voltage as shown in Figure 4. 8.
Figure 4. 8: Output power of the SOA as a function of the input voltage applied
The SOA was also tested for sufficient Optical Signal to Noise Ratio (OSNR)
performance to provide the optimum operating point for the first stage as shown in
Figure 4. 9.
71
Figure 4. 9: Optical Signal to Noise Ratio (OSNR) performance of first stage wavelength converter system
4.2.3 Second stage conversion and packet steering
An integrated SOA based MZI WC along with a fast tunable SGDBR laser is used as
the second stage for WC as described in [2]. The SGDBR laser is tuned to the
selected wavelength by the core node electronics through a fast switching board
similar to that used at the network ingress. Switching times for various input and
output wavelength combinations were measured and was found to be under 6nsecs for
all combinations. Packets entering the 2nd stage at the internal wavelength are
converted to the output wavelength selected on a per packet basis. A Fiber Bragg
Grating (FBG) based notch filter and circulator are used to reject the internal
72
wavelength and transmit packet payloads only at the output wavelength as shown in
Figure 4. 10.
MZI – WC2
λoutputλinternal
FBG @λinternal
Packet @ λoutputPayload @ λinternal
λinternal
drop
throughMZI – WC2
λoutputλinternal
FBG @λinternal
Packet @ λoutputPayload @ λinternal
λinternal
drop
through
Figure 4. 10: Mach Zehnder Interferometer based 2nd stage Wavelength Conversion. FBG – Fiber Bragg Grating
The 2nd stage WC is operated in the inverting mode to preserve the polarity of the
payload data bits and also serves to present a blank CW signal for label rewrite due to
the nature of the inverting operation.
4.2.4 Framing and idlers revisited
After payloads undergo two-stage WC, new framing information is rewritten onto the
packet stream before space switching is performed. A rewrite signal controlled by the
core node electronics is modulated onto the converted packet stream using an electro-
optic modulator (LiNbO3 modulator). It is critical to note that before rewrite is
completed, the optical payload stream does not present a constant average optical
power and therefore would result in transient distortions if amplifiers (such as
EDFAs) were used. Specially designed modulator drivers with low ac cutoff
frequency were used at the rewrite stage to ensure proper operation even without
73
adequate dc balance of the rewrite signal (constant high to pass the optical payloads).
Although locally generated idlers are used at the rewrite stage to fill interpacket gaps,
space switching or statistical DMUX of the packet stream results in additional void
creation at the output. External idlers controlled by the core node electronics are used
after space switching to ensure constant average output power at both the output ports
as shown in Figure 4. 11. An AWGR was used to perform actual space switching in
compliment to the fast tunable MZI WC.
AWGR out1
External IdlersVoids filled by external idlers
λ1 λ2 λ1 λ1
λ2
out2Internal Idlers
AWGR out1
External IdlersVoids filled by external idlers
λ1 λ2 λ1 λ1
λ2
out2Internal Idlers
Figure 4. 11: External Idler Insertion. AWGR - Arrayed WaveGuide Router
4.3 Experimental Performance Evaluation
In order to ensure proper operation of the core node optics, several physical layer
tests were performed on optical components. A schematic of the complete
implementation of the core node is shown in Figure 4. 12. The incoming packet
stream is kept in the optical forwarding plane and is passively delayed to match the
processing time of the core controller and subsequently fed to a two-stage wavelength
converter. The first WC stage removes the optical framing and the second stage
performs wavelength switching. An electro-optic modulator at the output of the two-
74
Figure 4. 12: Core node experimental setup
λ inte
rnal
CW
SOA
AO
WC
A W G R
λ 1co
re
. . . .
λ 2co
re
Idle
r W
ord
Gen
Q Q
OL,
Fra
me
eras
e
Dat
a in
to c
ore
λ in=
1548
.94n
m
Del
ay 1
λ inte
rnal
1551
.01n
m
λcore
out
cont
rol
λ inte
rnal
EOM
EOM
EOM
FBG
New
OL,
Fra
me
rew
rite
EDFA
Idle
r C
ontr
ol
1stSt
age
WC
2n
dSt
age
WC
Idle
r Gen
erat
ion
Rew
rite
λ id
ler 2
CW
λ id
ler 1
CW
Rx
Cor
e C
ontr
olle
r
OL,
Fra
me
eras
e
λcore
out
cont
rol
New
OL,
Fra
me
rew
rite
Elec
tron
ic C
ontr
ol
clk
data
Idle
r Con
trol
Opt
ical
ta
p
Ch1
Ch2
Idle
r for
C
h1
Idle
r for
C
h2
λ 1co
re =
155
5.86
nm
λ 2co
re =
156
0.71
nm
λ inte
rnal
CW
SOA
AO
WC
A W G R
λ 1co
re
. . . .
λ 2co
re
Idle
r W
ord
Gen
Q Q Q
OL,
Fra
me
eras
e
Dat
a in
to c
ore
λ in=
1548
.94n
m
Del
ay 1
λ inte
rnal
1551
.01n
m
λcore
out
cont
rol
λ inte
rnal
EOM
EOM
EOM
FBG
New
OL,
Fra
me
rew
rite
EDFA
Idle
r C
ontr
ol
1stSt
age
WC
2n
dSt
age
WC
Idle
r Gen
erat
ion
Rew
rite
λ id
ler 2
CW
λ id
ler 2
CW
λ id
ler 1
CW
λ id
ler 1
CW
Rx
Cor
e C
ontr
olle
r
OL,
Fra
me
eras
e
λcore
out
cont
rol
New
OL,
Fra
me
rew
rite
Elec
tron
ic C
ontr
ol
clk
data
Idle
r Con
trol
Opt
ical
ta
p
Ch1
Ch2
Idle
r for
C
h1
Idle
r for
C
h2
λ 1co
re =
155
5.86
nm
λ 2co
re =
156
0.71
nm
75
stage wavelength converter imprints new framing onto CW light surrounding the
converted optical packet payload, which is amplified and forwarded to separate space
ports using an Arrayed Wave Guide Router (AWGR). Optical packet headers and
packet frame delimiters are identified in the electronic core control plane by detecting
and processing the incoming optical packet stream. The control signals for framing
removal, wavelength switching, framing rewrite and idler insertion are all derived
from the core FPGA controller board which identifies individual optical packets,
extracts all-optical processing plane.
Scope traces of the packet processing within the core node are given in Figure 4. 13
(a) where incoming packets are detected and internal CW is aligned with the packet
payload at the input of the 1st stage WC.
idlers idlers
EOOPSOOP
Payload container
idlers idlers
EOOPSOOP
Payload container
Payload container
Figure 4. 13: 1st stage Wavelength Converter Scope traces (a) CW + packet (b) Frame erased wavelength converted payload
Figure 4. 13 (b). shows the wavelength converted payloads at internal wavelength
where framing has been erased while Figure 4. 14 (a) shows the payload after 2nd
stage conversion. The CW control based label erasure results in extremely good label
extinction ratio as can be seen in the scope trace. This ensures that previous label does
76
not corrupt the newly rewritten label and result in errored headers. Figure 4. 14 (b)
shows the optical packet after successful rewrite.
SOIP EOIPIP packetSOIP EOIPIP packet
Figure 4. 14: Scope traces after 2nd stage Wavelength Conversion (a) Before rewrite (b) After frame rewrite
4.3.1 Measurement test system and setup
The test configuration for layer-3 performance measurements over the entire optical
network implementation is shown in Figure 4. 15. Two separate IP streams with
different IP destination addresses were generated at the POS transmitter and adapted
at the ingress before being injected into the core. At the core, packet forwarding is
done based solely on the optical label generated at the ingress based on the IP
destination address of each packet.
77
Sm
artb
itsS
ON
ET
Test
er(R
x)
Hig
hS
peed
Ele
ctro
nic
Rou
ter
PO
SE
ncap
-su
latio
nE
gres
s #1
PO
SE
ncap
-su
latio
nE
gres
s #2
Egr
ess
Opt
ical
Rou
ter
#1
Egr
ess
Opt
ical
Rou
ter
#2
Sm
artb
itsS
ON
ET
Test
er(T
x)
PO
SD
ecap
-su
latio
n
Ingr
ess
Opt
ical
Rou
ter
OC
-48
Cor
eA
ll-O
ptic
alR
oute
rO
C-
48 OC
-48
IP F
LOW
1
IP F
LOW
2
(1)
(3)
(4)
(5)
(6)
(2)
Figure 4. 15: Traffic flow and Layer-3 measurement setup
78
4.3.2 Measured throughput performance
Figure 4. 16 shows the measured layer-3 packet throughput variation versus packet
size. The POS throughput of the electronic router (ER), a GSR 12000, was measured
through path 1-5-6 as shown in Figure 4. 15.
0 200 400 600 800 1000 1200 1400 16000
10
20
30
40
50
60
70
80
90
100
Thro
ughp
ut (%
)
Packet Size (bytes)
ER 1-5-6 Theoretical IE B2B 1-2-4-6 CDR 1-2-3-4-5-6 CDNR 1-2-3-4-5-6
Figure 4. 16: End-to-End IP throughput measurement and theoretical comparison (refer Figure 4.15)
A theoretical computation of the achievable throughput (% of available POS payload
bandwidth) accounts for the optical packet header and framing overheads and
Ingress-Egress back-to-back (IE B2B) throughput measurement shows that the edge
node adaptation closely matches the theoretical throughput for different packet sizes.
Throughput for Core Dynamic forwarding with or without Rewrite (CDR/CDNR)
shows that no significant throughput penalty is incurred both due to either wavelength
79
switching and conversion or label swapping. The 4% penalty measured without label
swapping is attributed to imperfect optimization of the two stage wavelength
converter for sufficient extinction ratio at the output.
4.3.3 Measured latency performance
Latency measurements were performed through the core to investigate the excess
time delay penalty due to core node switching as shown in Figure 4. 17.
0 200 400 600 800 1000 1200 1400 16000
5
10
15
20
25
30
35
40
Lat
ency
(use
c)
Packet Size (bytes)
IE-ER 1-2-4-5-6 CDNR 1-2-3-4-5-6 CDR 1-2-3-4-5-6 IEB2B 1-2-4-6 ER 1-5-6
Figure 4. 17: End-to-End latency vs. packet size (refer Figure 4.15)
Variation in packet latency with packet size is linear as expected due to the nature of
the simple lookup at the core node. The bulk of the latency seen in the system results
from the ER and the Ingress-Egress Adaptation and a comparison between latency
measured for the fully switching network system versus Ingress-Egress and
80
Electronic Router (IE-ER) path shows that only a minimal latency increase (max
0.79usec) is introduced by the core node. This indicates that optical packets may
undergo forwarding at multiple core nodes without significantly impacting the end-to-
end latency through the optical network.
4.4 Chapter Summary
All-optical packet forwarding based on optical header information was investigated in
this chapter. Basic sub-systems that go into building a core node were examined, and
their limitations and the influence on network performance was measured
experimentally and compared to theoretical limits. We experimentally demonstrate a
complete dynamic layer-3 (IP) forwarding AOLS network and present the first true
end-to-end layer-3 performance measurements. The agreement between the measured
throughput and the theoretically predicted is excellent and only show a penalty for
small packet sizes, which is ascribed to the electronic processing in the SONET layer
outside the AOLS network. The latency added is constant due to the fixed optical
path through the core node and measured to 0.79usec. It is negligible compared to
electronic switching and adaptation. Throughput penalty may be further reduced by
tightening control signaling through the core and improving optical rise/fall response
times. We successfully demonstrate an edge node adaptation that facilitates all-optical
forwarding at the core nodes.
81
References
[1] R. Doshi, M. L. Masanovic, and D. J. Blumenthal, "Demonstration of
regenerative any /spl lambda//sub in/ to any /spl lambda//sub out/ wavelength
conversion using a 2-stage all-optical wavelength converter consisting of a
XGM SOA-WC and InP monolithically-integrated widely-tunable MZI SOA-
WC," presented at Lasers and Electro-Optics Society, 2003. LEOS 2003. The
16th Annual Meeting of the IEEE,477-478 vol.2, 2003.
[2] V. Lal, M. Masanovic, D. Wolfson, G. Fish, C. Coldren, and D. J.
Blumenthal, "Monolithic widely tunable optical packet forwarding chip in InP
for all-optical label switching with 40 Gbps payloads and 10 Gbps labels,"
presented at Optical Communication, 2005. ECOC 2005. 31st European
Conference on,25-26 vol.6, 2005.
82
Chapter 5
Core Traffic Adaptation for
Statistical Multiplexing
In the previous two chapters, we investigated edge node adaptation that would be
necessary to support packet all-optical forwarding in the network core. A new
framing mechanism was devised and statistical demultiplexing at the network core
leveraging on this framing mechanism was proposed and successfully demonstrated.
In order to replace today’s electronic routers however, an efficient multiplexing
scheme is necessary at the high bit and packet rates that optics is capable of
supporting. Statistical multiplexing or aggregation of packets within the network core
relies on the fact that not all inputs require the use of the output resource at all times.
Buffering aids in statistical multiplexing by absorbing transients in shared capacity
demands. The absence of optical Random Access Memories (ORAM) severely limits
the forwarding throughput achievable and offsets many of the advantages in OPS. In
this chapter, we investigate the primary issue with implementing statistical
multiplexing namely contention handling and the techniques that improve the
performance of a core optical switch. Optical adaptation may be used to assist in
83
implementing contention handling and we look at strategies to deploy such adaptation
within the optical network.
5.1 Core node with contention resolution
The core forwarding model proposed and implemented thus far looks at packet
forwarding in the optical domain from the perspective of the optical packet stream
itself. It does not take into account the possibility of a second data stream requiring
shared access to the same output ports at the same time. When such multiple data
streams look to share the same set of output ports over time, a technique to multiplex
packets from the different inputs by handling contention becomes necessary as shown
in Figure 5. 1.
Multiple Inputs
Multiple Outputs
CR
CR
CR
CR
Contention Resolution
Multiple Inputs
Multiple Outputs
CR
CR
CR
CR
Contention Resolution
Figure 5. 1: Contention resolution during statistical MUX. CR- Contention Resolution
5.1.1 Statistical Multiplexing and contention
Statistical multiplexing is the aggregation of multiple input channels based on the
probability of need to access a shared output channel and can result in an efficient use
of a shared resource. Multiplexing packets from multiple inputs to the same output is
possible as long as the aggregate capacity demand placed by all the inputs does not
84
exceed the capacity of the shared output port for an extended period of time.
Contention occurs when two or more input ports simultaneously request the use of the
same output port that cannot be shared. At low input utilizations, contention rarely
occurs and packet loss rates are low and usually below tolerance limits. However, as
the aggregate input utilization normalized to the output capacity, tends towards to 1,
packet loss rates increase and contention resolution schemes become necessary.
Successful resolution requires 1) prior knowledge of contention with time to react and
resolve it, 2) an intelligence capable of making decisions on how to resolve
contention and 3) a resolution mechanism to implement contention resolution. Arrival
information through detection of packets by the core electronics may be used both to
detect contention and to decide on the path to resolving contention. In [1] we
demonstrate an all-optical contention resolution scheme with a first come first serve
decision rule for variable length 40 Gbps payloads.
5.1.2 Contention handling schemes
Traditionally, contention may be resolved in either the time domain by use of a buffer
to delay the transmission of the contending packet or in the space domain by steering
the contending packet to a different output port. While optics suffers from the
drawback of the lack of availability of ORAMs, access to an additional dimension
namely the wavelength can be exploited for contention resolution. Fast tunable
wavelength conversion may be used to provide access to a shared set of wavelengths
within a delay loop to provide suitable resolution at packet rates.
85
5.2 Optical Buffering techniques
Optical buffering provides a traditional approach to contention resolution in ensuring
highly efficient optical switches. However, while several alternate buffering
techniques such as slow light in semiconductor nanostructures [2] and arrayed-rod
photonic crystals [3] are being explored, currently fiber based delays structures are
the only available technology to implement optical buffering. Such delay lines offer
fixed and finite amount of delay and may be either used in a single stage or as
multiple stages. Delay based buffers may also be classified as feedback or feed-
forward buffers based on the nature in which they are connected [4].
5.2.1 Feed-forward buffering schemes
Figure 5. 2 shows a typical structure of a feed-forward fiber delay line based buffer.
Fiber delay lines of logarithmically increasing amount of delay are usually used
together and the amount of delay is selected by either traversing a series of delay
elements or bypassing them.
Packet in Delayed packet out
2 x 2 switchFiber delay lines
Packet in Delayed packet out
2 x 2 switchFiber delay lines
Figure 5. 2: Feed-forward delay structure
In feed-forward delay structures, access to the shared output is guaranteed to the
packet once a delay is set for it. Consequently implementing a packet priority
mechanism is difficult to implement. Nonetheless, several feed-forward architectures
86
have proposed including the Cascaded Optical Delay-lines (COD), Switched Fiber
Delay-lines (SDL) and the Logarithmic Delay-Line Switch.
5.2.2 Feedback buffering
Feedback buffering uses a fiber delay loop between one input–output pair of the
optical switch as shown in Figure 5. 3. Contending packets may be stored in the fiber
delay loop for a fixed period of time before contending for access to the desired
output port at a later point in time.
Packets in Delayed packet out
Switch element
Fiber delay lines
Packets in Delayed packet out
Switch element
Fiber delay lines
Figure 5. 3: Feedback delay buffering
Feedback buffering however requires that the length of the feedback loop is large
enough to store the largest packet completely to avoid data corruption. By using fixed
time slots for packets, feedback buffering can efficiently resolve contention.
5.3 Compression/Decompression for contention resolution
Compression takes advantage of the transparency of optics and may be used as a way
to speed up packet payloads so that they occupy a smaller time slot. The time slot
87
selected must be the size of the largest variable sized packet at the highest
compressed bit rate. Compression of all packets at the highest bit rate requires
padding with non-zero data to ensure that the entire time slot is filled for dc balance.
This however can be cumbersome and may require tedious synchronization between
the compressed packet and the padding. By using variable compression ratios to
compress packets of different sizes, each packet may be adjusted to occupy the entire
fixed time slot and therefore eliminate the need for padding. An optical adaptation
mechanism that would provide such compression/decompression between variable
byte sized packets and a fixed time slot in addition to providing bit rate speedup
would therefore be useful in simplifying implementation of contention resolution
schemes.
5.3.1 Network-wide deployment
When compression /decompression are deployed at the network edges, incoming
traffic is aggregated into fixed sized higher bit rate compressed packets before being
injected into the network. At the egress, optical payloads are decompressed back to
the base rate of the original payload. Packets in the network are therefore of fixed
temporal duration throughout the network as shown in Figure 5. 4.
88
1x
1x
1x
1x
8x
2x4x
8x
16x
2x
16x
Ingress CO
M
Ingress
CO
M
EgressDC
OM
Egress
DC
OM
Fixed base rate, variable
temporal sizes
Variable compressed rate, fixed temporal
sizes
Fixed base rate, variable
temporal sizes
1x
1x
1x
1x
8x
2x4x
8x
16x
2x
16x
Ingress CO
M
Ingress
CO
M
EgressDC
OM
EgressDC
OM
Egress
DC
OM
Fixed base rate, variable
temporal sizes
Variable compressed rate, fixed temporal
sizes
Fixed base rate, variable
temporal sizes
Figure 5. 4: Compression/Decompression deployed at the network edges
However, the link bandwidth utilization for packets smaller than the largest packet
goes down as smaller packets are padded to occupy the temporal time slot of the
largest packet at the compressed rate as shown in Figure 5. 5. For an IMIX traffic
distribution of 1500,560 and 40 byte packets in a 7:4:1 ratio, the link utilization is
calculated to be only 26.5% when using discretely variable compression ratio
compression scheme. Network wide deployment of compression/ decompression has
the advantage of reducing the need for compression/decompression at every core
node. It is feasible when a large number of core nodes are present and the link
bandwidth has been over-provisioned throughout the network. Compression ratio and
original packet length information must however be embedded within the optical
header in order to successfully decompress at the egress nodes.
89
Util without padding %
0.00
10.00
20.00
30.00
40.00
50.00
60.00
70.00
80.00
90.00
100.00
0 500 1000 1500
Util without padding %
Figure 5. 5: % available link utilization Network scale ComDecom vs. packet size in bytes
5.3.2 Intra-node deployment
Alternately, compression/ decompression may be applied at the input and output of
every core node as shown in Figure 5. 6. This deployment scheme provides the core
node switching benefits of compressing packets to fixed temporal sizes and higher bit
rates while maintaining 100% link utilization. This approach also has the advantage
that high speed packets do not have to be transported on network links. Variable
length packets entering each core node are compressed to occupy a fixed time slot
large enough to hold the largest length packet at the highest compressed bit rate
before entering the switching element. Once a packet has been successfully switched,
it is decompressed back to base rate before reframing and exiting the core node. Since
a single controller can be used to control compression, switching and decompressing
of each packet, knowledge of the packet resides within the controller for the duration
90
while the packet resides in the core node. Control signals may therefore be
synchronized for each packet within the node.
.
.
.Pack
et
Com
pres
sion
Pack
et
De-
com
pres
sion
.
.
.Synch & Traffic shaping
Switch element
Packet ContentionResolution
Output Traffic Merge
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Feedback buffering
Feed-forward buffering
Fixed base rate,variable sized packets
Variable compressed rate,fixed sized packets
Fixed base rate,variable sized packets
.
.
.Pack
et
Com
pres
sion
Pack
et
De-
com
pres
sion
.
.
.Synch & Traffic shaping
Switch element
Packet ContentionResolution
Output Traffic Merge
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Feedback buffering
Feed-forward buffering
.
.
.Pack
et
Com
pres
sion
Pack
et
De-
com
pres
sion
.
.
.Synch & Traffic shaping
Switch element
Packet ContentionResolution
Output Traffic Merge
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Feedback buffering
Feed-forward buffering
Fixed base rate,variable sized packets
Variable compressed rate,fixed sized packets
Fixed base rate,variable sized packets
Figure 5. 6: Intra-node deployment of Compression/decompression at the core
Deployment of core-centric compression/decompression places stricter requirement
on the compression/decompression technique. These increased constraints come from
when packets are decompressed to base rate; they occupy a larger temporal size than
when compressed. Compressed packets would therefore be required to have adequate
spacing between them before entering the decompressor to avoid the overlapping at
the base rate (a condition called time-out that is discussed in more detail in Chapter
7). The core control plane may buffer the compressed packets after switching to
ensure that they are spaced out in time to avoid the packet overlap issue at the
decompressor.
5.3.3 Compression/Decompression process
Compression and decompression involve temporal rearrangement of bits within the
packet and a translation in bit rate to or from a higher bit rate. In the case of both
91
processes, control signals required must have sufficient rise/fall times and granularity
of control to prevent corruption of packet bits. In addition to this requirement,
compression and decompression need suitable modifications to the bits before and
after the processes to facilitate the temporal rearrangement. Figure 5. 7 shows the
steps required before packet compression and decompression may be performed.
CR CR -- Clock RecoveryClock Recovery
CRCR DriveDrive
NRZ NRZ -- to to -- RZRZ CompressionCompression DecompresDecompres--sionsion RZ RZ -- to to -- NRZNRZ
Optical network Optical network or core or core
switching switching elementelement
Incoming low bit rate NRZ packets RZ packets Compressed RZ
packetsOutgoing low bit rate NRZ packets
CR CR -- Clock RecoveryClock Recovery
CRCR DriveDrive
NRZ NRZ -- to to -- RZRZ CompressionCompression DecompresDecompres--sionsion RZ RZ -- to to -- NRZNRZ
Optical network Optical network or core or core
switching switching elementelement
Incoming low bit rate NRZ packets RZ packets Compressed RZ
packetsOutgoing low bit rate NRZ packets
Figure 5. 7: Compression Decompression process detail
Incoming packets at lower bit rates usually use the Non-Return-to-Zero (NRZ) bit
representation format. Compression of packets requires a reduction in the effective bit
period in representing packet information. Conversion of packets from NRZ to
Return-to-Zero (RZ) bit format is one way of achieving this and may be done before
the compression subsystem. Recovery of bit boundary information is important to
performing NRZ-to-RZ conversion and a suitably time aligned signal from a packet
clock recovery system may be used. One approach is to use an electro-optic gating
modulator such as an EAM with a suitable transfer function to carve RZ pulses. After
decompression, it is desirable to represent bits within the packet back in NRZ format
and several techniques have been previously demonstrated to perform this
conversion. The quality of the RZ pulses entering the compressor is critical in
92
determining the quality of the compressed packets. The pulse width must be less than
one bit slot at the compressed data rate while the extinction ratio (Figure 5. 8) must be
sufficiently high (typically >25dB) for successful compression.
High extinction
Small pulse width
t
High extinction
Small pulse width
t
Figure 5. 8: RZ pulse quality requirements for compression
5.4 Chapter Summary
Contention resolution is necessary at the network core to perform low loss statistical
multiplexing of optical packets. We propose the use of bit rate transparency of all-
optical forwarding schemes to achieve adaptation of variable byte size packets into
fixed time slots, thereby making it easier to forward packets using limited contention
resolution resources by trading in bandwidth. Use of fixed sized packets however
requires slot synchronization for efficient forwarding and schemes have been
proposed to implement such synchronization. Two proposed network deployment
strategies were proposed for use of compression and decompression to implement
optical adaptation, namely the intra-node deployment and the network-wide
deployment. The intra-node deployment is chosen as it decouples the link utilization
from adaptation to aid in efficient core node forwarding. Moreover, this scheme relies
93
on a centralized control plane for compression and decompression making it easier to
recover the original payload from high bit rate compressed packets.
94
References
[1] S. Rangarajan, H. Zhaoyang, L. Rau, and D. J. Blumenthal, "All-optical
contention resolution with wavelength conversion for asynchronous variable-length
40 Gb/s optical packets," Photonics Technology Letters, IEEE, vol. 16, pp. 689-691,
2004.
[2] C. J. Chang-Hasnain, K. Pei-cheng, K. Jungho, and C. Shun-lien, "Variable
optical buffer using slow light in semiconductor nanostructures," Proceedings of the
IEEE, vol. 91, pp. 1884-1897, 2003.
[3] M. Tokushima, "Experimental demonstration of waveguides in arrayed-rod
photonic crystals for integrated optical buffers," presented at Optical Fiber
Communication Conference, 2005. Technical Digest. OFC/NFOEC,3 pp. Vol. 3,
2005.
[4] D. K. Hunter, M. C. Chia, and I. Andonovic, "Buffering in optical packet
switches," Lightwave Technology, Journal of, vol. 16, pp. 2081-2094, 1998.
95
Chapter 6
Optical Packet Compression
Statistical multiplexing of optical packets, along with statistical demultiplexing
shown in chapter 4, facilitates packet forwarding at the core from multiple inputs to
multiple outputs. Contention resolution is required to efficiently perform statistical
multiplexing. One way to implement contention resolution is by the use of fixed line
delays as buffering elements. However, due to the inherent nature of today’s delay
line buffers, handling variable sized packets requires tradeoffs in utilization and
efficiency. Optical compression is one way of adapting packets to suit fixed line delay
buffers and implement contention resolution. Compression temporally squeezes each
packet thereby reducing the service time required to forward the packet, be it access
to output ports or storage in optical delay line buffers. Unlike conventional schemes
that slice the optical spectrum using wavelength division multiplexing to tap the large
available optical bandwidth, packet compression aims to use separate high bandwidth
wavelengths to provide ultra-high speed channels. Previous techniques proposed to
implement packet compression have not been scalable in packet size handling
capability. In chapter 5, we proposed a network architecture suitable for deploying
packet compression/decompression to improve forwarding performance in a multiple
input/ output node. In order to enable statistical multiplexing, requirements on a
96
packet compressor are identified and a suitable technique is proposed and
demonstrated. The performance of the compression technique is then studied both on
its versatility and the bit level packet quality after compression.
6.1 Compression – State of the art today
To date, several techniques have been proposed and demonstrated to implement
packet compression. This section details the state of art in compression technology,
and the challenges/drawbacks in practical deployment of such techniques. Each
technique detailed in this section requires that the compression process maintains the
bit order within the packet after compression thereby requiring distinct manipulation
of every bit in the packet at the compressed rate. A second requirement of all
proposed techniques is the availability of several compression control signals with
rise and fall times and control granularity of the compressed bit rate. Such a
requirement negates the advantage of all optical packet forwarding where flow of
high speed traffic in the optical domain is controlled by low data rate electronics.
Compression schemes thus far require several block elements to successfully process
packets and the complexity of such compressors scales with the maximum packet
length. With the growth in average packet length in the internet [1], it is desirable to
have a compression technique whose complexity has a low dependence on the
maximum input packet length. Compression ratio achievable in current schemes is
fixed for all packet lengths and permanent setup changes are usually required to
achieve different compression ratios. Furthermore, none of the current techniques for
97
compression have studied or proposed a way of handling variable length input
packets. Given the packet size distribution found in today’s internet [1], the ability of
a compressor to handle a wide range of input packet lengths becomes essential.
6.1.1 Feed forward delay line approach
One proposed approach to packet compression uses parallel delay line structures. In
this approach, splitting the incoming signal creates multiple copies of the packet.
Each copy of the packet is then appropriately shifted by a distinct delay amount to
align a specific bit of the incoming packet to its corresponding compressed bit slot.
The delayed bit sequences are multiplexed back in time using a combiner and a high
speed gate is used to select the appropriate compressed bits. Active sections may be
used between sections to compensate for the 3dB losses due to power splitting.
Gating may be done either at the end of each section or the output of the compressor.
The general structure of such a compressor is detailed in Figure 6. 1 (a). Figure 6. 1
(b) shows the principle of operation of the feed-forward delay compression approach.
Successful isolation of the compressed bits relies on the quality of the control signals.
Each compression element requires a unique control signal with compressed rate
precision and control granularity for error- free compression of the entire packet. The
number of cascaded compression elements required increases logarithmically with the
input packet length and is given by M = log2(N) where N is the number of bits in the
input packet. Moreover, each of the M delay elements must have an accuracy of a
fraction of the compressed bit slot.
98
…
(T-t) 2(T-t) 2N(T-t)Optical packet
IN
Compressed packet
OUT
- Coupler or switch element
…
(T-t) 2(T-t) 2N(T-t)Optical packet
IN
Compressed packet
OUT
- Coupler or switch element
(a) 21 43 65
21 43 65
87
87
21 43 65 87
Gate Gate
21 43 65
21 43 65
87
87
21 43 65 87
Gate Gate (b)
Figure 6. 1: (a) Structure of a feed forward delay compressor, (b) Working Principle
In [2] for example, this requirement for a 9 stage compressor with a maximum packet
size of 512 bits, would require a cleaving/splicing accuracy of ±200um (1psec) on
multiple delay lines of up to 50 m or a 0.0004% accuracy. Control over several delay
lines to such a precision is a design challenge both in implementation and stability
over environmental changes such as temperature. Such a scheme would have a
maximum packet size of less than (K-1) where K is the compression ratio due to the
inherent nature of compression. A similar approach is also demonstrated in [3] uses
multiple Optical Delay Line Lattices (ODLLs) to increase the maximum packet size
that can be handled. However, this scheme also suffers from increased complexity,
instability from environmental changes and stringent requirements imposed on the
compression control signals. The compression ratio achievable using this method is
fixed based on the delay values selected. Variable compression ratio implementation
would therefore require permanent setup changes in the scheme.
99
6.1.2 Spectral Slicing approach
A second approach [4] to compression uses the delay and combine technique as well
but utilizes the wavelength dimension to implement it. Incoming packets are copied
to multiple wavelengths using super continuum generation followed by spectral
slicing. Each wavelength copy is then separated, delayed appropriately and
multiplexed back together to perform timing alignment necessary for compression
using Arrayed Wave Guide Router (AWGR) and delay lines. The compressed packet
is returned to a single wavelength by spectral slicing of the generated super
continuum. Figure 6. 2 shows the operation of this approach.
1 2 3 4 λin
21 43λ1 ... λ4
1 2λ2
3λ3
4λ4λ1
21 43λout
1 2 3 4 λin
21 43λ1 ... λ4
1 2λ2
3λ3
4λ4λ1
21 43λout
Figure 6. 2: Spectral slicing approach - Principle of operation
Decompression may be performed using the same setup in the reverse direction and
using a narrow gating window with base rate repetition rate. While the method
reduces power losses due to splitting and multiplexing by using an AWGR and the
wavelength domain, the complexity of the system increases rapidly with the
maximum packet size. The number of wavelengths required in this approach equals
the number of bits in the packet. A second method proposed in [5] also employs the
wavelength domain and uses linearly frequency chirped pulses to obtain time-to-
100
wavelength conversion. However, 1.15 Km of fiber was used to compress a packet of
only 4 bits. Compression/decompression of large packets is therefore impractical
using either technique. Compression control signals are required to be of compressed
rate accuracy and control granularity. Implementation of variable compression ratio
would require a complete change of the delays and wavelengths used i.e. permanent
hardware changes.
6.1.3 Loop based approach
The oldest approach presented thus far [6] uses recirculating loops to align and store
compressed packet bits until the entire packet has been compressed. Bits are switched
into a recirculating loop whose loop length is adjusted to be one compressed bit
shorter than the base rate bit slot. Once a packet has been compressed, it is gated out
of the loop. Figure 6. 3 shows the principle of operation of the recirculating loop
based approach.
21 43 65 87
21 43321 21 65 87765 65
Gate
12 3 4 567 8
21 43 65 8721 43 65 87
21 43321 21 65 87765 6521 43321 21 65 87765 65
Gate
12 3 4 567 8
Gate
12 3 4 567 8
Figure 6. 3: Loop based compression - Principle of operation
The maximum number of bits is limited to the (K-1) where K is the compression
factor. The maximum packet size is also limited by ASE buildup in the loop, as the
number of recirculations required equals the number of bits in the packet. [7]
experimentally demonstrates the recirculating compression scheme and analyzes the
101
OSNR performance with increasing packet size. OSNR degradation limits the packet
size to below 100 bits, a size that is impractical for most network applications. The
number of compression elements required however does not increase rapidly with
increasing input packet size. This scheme may therefore be suitable for packet
compression if modified appropriately. However, in order to get a clean compressed
packet, a compression control signal with compressed rate precision and granularity is
required.
A compression mechanism chosen must be able to handle packet sizes of at
least 1500 bytes (12000 bits). To date, proposed and/or demonstrated schemes have
been able to handle a maximum of 100 bits with no practical possibility of scaling to
larger packet sizes. Any compression scheme proposed must also facilitate a viable
decompression mechanism without requiring compressed rate electronics to identify
and control compressed packets.
6.2 Network deployment of packet compression
Adaptation of packets for efficient packet forwarding at a multiple input/output core
node was analyzed in Chapter 5. Packet compression may be performed in several
ways based on the nature of the compressed packet at the compressor output. A study
of the deployment of packet compression/decompression is necessary to identify a
specific compression scheme and requirements placed on it. Successful end-to-end
forwarding of packets depends on the decompression scheme’s ability to identify
compressed rate packets and decompress each packet to its original temporal size
102
before it reaches the edge adaptation and exiting the optical network. The primary
requirement of any compression scheme is therefore to output compressed packets
that not only increase the core forwarding efficiency but also are decompressable
without bit errors for any input packet length.
6.2.1 Intra-node compression/decompression
Optical compression and decompression may be used to provide packet adaptation at
the core. However, payload adaptation may still be necessary at the network edge to
perform error free compression and decompression at the core. Figure 6. 4 shows the
adaptation required for efficient forwarding at the core using compression/
decompression. However, at the core, optical compression increases the payload bit
Com
p
Dec
omp
Pkt framing and core adaptation
INGRESS
EGRESS
Pkt framing and core adaptation
Base rate pkts
Compressed rate pkts
CORECORE
CORE
All Optical Network
Com
p
Dec
omp
Com
p
Dec
omp
Com
p
Dec
omp
Pkt framing and core adaptation
INGRESSINGRESS
EGRESSEGRESS
Pkt framing and core adaptation
Base rate pkts
Compressed rate pkts
CORECORE
CORE
All Optical Network
Figure 6. 4: Intra node deployment of compression and decompression
103
size before exiting the core. While compression/decompression are intended to
increase core forwarding efficiency and enable the use of short fixed size buffers,
they are not necessarily bandwidth efficient. Intra-node compression/decompression
deployment separates the high throughput fiber links between nodes from the
switching process at the core. This allows for maintaining high link utilization
regardless of the compression scheme that is tailored to enable and maximize
forwarding at the NxN core.
IP header extraction
Optical packet
framing
Optical header lookup
Payload (de)encoding
for core adaptation
Optical modulation
on λselect
Optical packets
Ingress node
IP header extractionIP header extraction
Optical packet
framing
Optical header lookup
Payload (de)encoding
for core adaptation
Optical modulation
on λselect
Optical packets
Ingress node
Figure 6. 5: Edge ingress node functionality– electronic adaptation
Optical packets IP packetsIP header
extraction
Optical payload
deframing
Payload decoding for
core adaptation
Photo detect and CDR
Egress node
Optical packets IP packetsIP header
extractionIP header extraction
Optical payload
deframing
Payload decoding for
core adaptation
Photo detect and CDR
Egress node
Figure 6. 6: Edge egress functionality – electronic adaptation
Figure 6. 5 and Figure 6. 6 show the electronic adaptation necessary on the packet
payload at the ingress and egress for successful packet adaptation at the core. At the
ingress, incoming IP packets are examined to lookup an optical header with
104
appropriate forwarding information. Each IP packet is individually framed into
optical packets with payload framing and appropriate header and packet framing. The
payload then undergoes electronic processing required to facilitate error free
compression/decompression process at the core. Optical packets are now ready to be
modulated on a selected wavelength and exit the ingress node. At the egress, the
reverse process takes place. For successful core node packet adaptation, the payload
bit sequence must not be altered after processing for compression/decompression.
Subsequently, payload encoding is the last step at the ingress before packet
transmission and the first step at the egress after packet detection.
An electronic control plane at the core is used to identify the time of arrival and
payload length of the each incoming optical packet. Control signals generated by the
electronic circuitry are used for compression of payloads before packet forwarding.
.
.
.Pack
et
Com
pres
sion
Pack
et
De-
com
pres
sion
.
.
.Synch & Traffic shaping
Switch element
Packet ContentionResolution
Output Traffic Merge
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Feedback buffering
Feed-forward buffering
Fixed base rate,variable sized packets
Variable compressed rate,fixed sized packets
Fixed base rate,variable sized packets
.
.
.Pack
et
Com
pres
sion
Pack
et
De-
com
pres
sion
.
.
.Synch & Traffic shaping
Switch element
Packet ContentionResolution
Output Traffic Merge
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Feedback buffering
Feed-forward buffering
Fixed base rate,variable sized packets
Variable compressed rate,fixed sized packets
Fixed base rate,variable sized packets
Figure 6. 7: Intra-node deployment of Compression/decompression at the core
After successful forwarding, packet traffic is merged and decompressed to base rate
before transmission out of the node as shown in Figure 6. 7. Adaptation of variable
105
length packets entering the core node into fixed sized payloads enables the efficient
use of short fixed length delay lines to buffer packets contending for an output port.
6.2.2 Packet compression classification
To realize buffering using current state of art technology, adaptation of variable
length packets to fixed time slots may be used before statistical multiplexing.
Variable length packets may be fragmented or compressed to adapt from variable
length to fixed sized packets. Fragmentation of packets before forwarding requires
fragment IDs to keep track of packet fragments. Independent forwarding of packet
fragments can also lead to disordering and a higher probability of error due to header
corruption. Another approach to perform this adaptation is packet compression.
Compression and decompression of packets to adapt variable length packets into
fixed sized cells may be achieved in one of several approaches. The compression ratio
may be either fixed for all input packet lengths, or varied discretely or continuously
on a per packet basis. Fixed compression ratio results in the same variance of
compressed packet sizes as that of the input packet sizes as shown in Figure 6. 8.
Packets may be padded to the maximum packet size after compression to reach a
fixed size. A compression scheme implementing constant compression ratio is
difficult to achieve both for fixed and variable input packet lengths as it requires bit
level manipulation throughput the packet.
106
CompressorVariable length packets@ base rate
Variable length packets@ compressed rate
4x 4x 4x
CompressorVariable length packets@ base rate
Variable length packets@ compressed rate
4x 4x 4x
Figure 6. 8: Fixed compression ratio compression – padding to achieve fixed temporal size
In order for packets of all lengths to occupy the same compressed time slot without
padding, continuously varying the compression ratio based on the input packet length
is required. However, such fine control over the compressor is hard to implement and
therefore a discretely variable compression ratio may be used as shown in Figure 6. 9
CompressorVariable length packets@ base rate
Fixed temporal size packets@ compressed rate
2x 1x 4x
CompressorCompressorVariable length packets@ base rate
Fixed temporal size packets@ compressed rate
2x 1x 4x
Figure 6. 9: Variable and discretely variable compression ratio compression
To achieve variable length to fixed size packet adaptation using a compression
scheme with discretely variable compression ratio, a small amount of padding is
required. This limited control over compression ratio allows for higher input
utilization as such a compressor modifies it’s processing for each input packet length,
thereby achieving improved performance.
6.2.3 Requirement of a practical compressor
A compressor/decompressor used to adapt packets from/to variable length IP packets
to/from fixed sized cells requires a compression ratio control, phase aligned clock,
and gating control signals for the compression elements as shown in Figure 6. 10.
107
Gate control Signals for (de)compression elements
Compression Ratio Control
Freq, phase aligned base
rate clock for pulse carving and compression
Data @ compressed/ base rate
Packet Compressor
/Decompressor
(De)Compressed packetstream @ base/ compressed
bit rate
Gate control Signals for (de)compression elements
Compression Ratio Control
Freq, phase aligned base
rate clock for pulse carving and compression
Data @ compressed/ base rate
Packet Compressor
/Decompressor
(De)Compressed packetstream @ base/ compressed
bit rate
Figure 6. 10: Generic structure of compressor/decompressor
Based on the network architecture proposed and studied, several requirements must
be satisfied by such a compressor/decompressor in order to successfully perform
adaptation of variable sized packets to fixed time slots.
• The scheme must support a maximum packet length of 1500 bytes (12000
bits) and a packet length variation from 40 bytes (320 bits ) to 1500 bytes.
• Discretely variable compression ratio changing on a per packet basis must be
achievable without having to implement permanent hardware changes.
• Control signals and clocks necessary for the compressor must be few and
generated asynchronously based solely on the packet timing information. The
number of signals required must not scale directly with the maximum input
packet length.
108
• Any control signals required for compression must be of base rate precision
and granularity in order to be viable for generation by the core node
electronics.
• Compression must take place in real time in order to support the maximum
line rate at the input.
• Any compression scheme selected must be scalable both in terms of
maximum input packet length handling capability and highest compressed bit
rate achievable for a given base rate.
• The proposed compressor technology must be robust against environmental
stress such as temperature fluctuations.
6.2.4 Approaches to compression implementation
Compression of a packet involves the temporal rearrangement of the packet bits to
reduce the footprint of the packet in time. However, such rearrangement is not
required to preserve the bit order of the packet so long as the original packet
information may be successfully reconstructed either by simple processing of the
compressed packet or if the decompression process reverts the packet bit order to its
original form. Such a relaxation of the constraint on the packet bit order reduces the
processing required to achieve compression drastically as individual bit level shifting
is no longer necessary. In the particular case of an all optical label switched network,
the optical payload that undergoes compression and decompression is not examined
until the packet reaches the egress and is ready to exit the network. Compression
schemes proposed/demonstrated thus far strive to maintain the bit order after
109
compression. In such a bit serial approach to compression, each bit must individually
be shifted in time by a distinct delay period to the corresponding bit slot at the
compressed bit rate as shown in Figure 6. 11. Such manipulation therefore requires
compressed bit rate precision control for every bit in the packet and subsequently
does not scale well with input packet length.
21 43 65 87
87654321
Compression Decompression
21 43 65 8721 43 65 87
87654321 877665544332211
Compression Decompression
Figure 6. 11: Bit- serial compression
A second approach to compression is shown in Figure 6. 12 where each packet may
be viewed as a number of contiguous fixed sized virtual bins. While the order of the
bits after compression does not follow that of the base rate packet, the number of
control signals required to perform compression/decompression does not scale
directly with the number of bits in the packet. Such an approach can handle both large
and small packets as discretely variable compression ratio is achieved by controlling
the number of bins included in the compression process. Each bin is buffered for a
multiple of a fixed delay to form the compressed packet. The compression process
takes place as the incoming packet is loaded into the compressor and the compressed
packet becomes available while the end of the packet is being loaded. The scheme
therefore imposes little or no latency to the packet. At decompression, the compressed
packet is sampled to demultiplex bits from each bin to reconstruct the base rate packet
with the correct order of bits.
110
40-1500 bytesbin size
1 bin size - 1 compressed bit slot1 bin size - 1
compressed bit slot
1 bin size - 1 compressed bit slot
1 bin size - 1 compressed bit slot
1 bin size - 1 compressed bit slot
1 bin size + n bits
Compressed packet
40-1500 bytesbin size
1 bin size - 1 compressed bit slot1 bin size - 1
compressed bit slot
1 bin size - 1 compressed bit slot
1 bin size - 1 compressed bit slot
1 bin size - 1 compressed bit slot
1 bin size + n bits
Compressed packet Figure 6. 12: Compression Principle - packet fold-in approach
Decompression is also performed while the packet is loaded and therefore imposes
little or no additional latency. The highest compressed bit rate is determined only by
the NRZ to RZ conversion process and the pulse width of the bits in the packet. The
control signal precision has minimal impact on the maximum compressed bit rate.
The fold-in compression/decompression approach is therefore scalable both in bit rate
and in the maximum input packet size and promises to satisfy the relaxed control
signal requirements of a practical packet compressor/decompressor.
6.3 Fold-in packet compression
Two approaches of implementing the fold-in compression scheme are proposed and
studied, namely the linear time delay compression and the loop based compression.
111
The loop based approach is selected for its scalability, ease of implementation and
compactness.
6.3.1 Linear time delay compression
IngressSOA-WC
NRZ to RZ
AWGR
AWGR
Super--continuum or λ1,λ2,λ3,λ4 CW
Envelope delay
Compression/ Fold-in delay
4 by 1 coupler
4 by 1 coupler
Fiber WC
λ0
λ0
λ1- λ4
λingress
λingress
λ1- λ4 λ1λ2
λ3
λ4
λ1λ2
λ3
λ4
Compressed packet on
Optical packet payload
IngressSOA-WC
NRZ to RZ
AWGR
AWGR
Super--continuum or λ1,λ2,λ3,λ4 CW
Envelope delay
Compression/ Fold-in delay
4 by 1 coupler
4 by 1 coupler
Fiber WC
λ0
λ0
λ1- λ4
λingress
λingress
λ1- λ4 λ1λ2
λ3
λ4
λ1λ2
λ3
λ4
Compressed packet on
Optical packet payload
Figure 6. 13: Linear time delay implementation of fold-in compression
Figure 6. 13 shows the principle of operation of the linear time delay approach.
Different parts of the incoming packet are wavelength converted to separate
wavelengths using a Semiconductor Optical Amplifier (SOA) based Cross Gain
Modulated (XGM) wavelength conversion. The packet then undergoes NRZ to RZ
conversion to reduce the pulse width of the bits. An Arrayed Wave Guide Router
(AWGR) is used to spatially separate sections of the packet and each section is
appropriately delayed before being multiplexed into a single output. A high speed
fiber wavelength converter is used to convert the compressed packet into a single
outgoing wavelength. Control signals required to turn on/off the separate CW’s are of
base rate precision. While the compressed packets have desired sharp edges, this
approach requires a multi-wavelength source and a high speed wavelength converter.
Polarization dependence of the high speed wavelength converter (WC) can also result
112
in degraded compressed packet bit quality. The number of delays with picosecond
accuracy required for compression scales directly with the input packet size and the
compression ratio required. The linear approach is therefore not very scalable in
terms of compressed bit rate and input packet size.
6.3.2 Loop based compression
A second approach proposed uses a fiber loop to provide the delay for each virtual
bin of the incoming packet. Figure 6. 14 shows the principle of operation of the loop
based implementation. Virtual bins of the incoming packet are stored in an
accumulative buffer and bit interleaved with adjacent virtual bins. The bit-interleaved
bin is then stored in the accumulative buffer to be time delayed appropriately and
ready for bit interleaving with the next incoming virtual bin. The compressor gating
outputs only the compressed packet when all incoming virtual bins have been
interleaved. The simple structure of the compressor requires only two simple control
signals, one to control and flush the accumulative buffer and another to control the
compressor gating after each packet has been compressed.
113
Bit levelbin interleaver
Accumulative Buffer
Compressor Gating
Base rate Compressed rate
virtual bins
Uncompressed packets
Compressed packets
Bit levelbin interleaver
Bit levelbin interleaver
Accumulative Buffer
Accumulative Buffer
Compressor Gating
Compressor Gating
Base rate Compressed rate
virtual bins
Uncompressed packets
Compressed packets
Figure 6. 14: Loop based compression principle
Compressed rate precision and granularity is not required on the control signals.
Figure 6. 15 shows an example of packet compression using the loop based
implementation. The size of the loop is chosen to be one compressed bit shorter than
the compressed packet time slot. After undergoing NRZ to RZ conversion, the packet
enters the loop. Each virtual bin of the incoming packet is distinguished by its
separate dotted line in the figure. Each section of the packet is multiplexed into a
fixed time slot with each loop trip. At the end of the fourth loop trip, all sections of
the incoming packet have been successfully multiplexed into a fixed time slot and are
gated to output only the compressed packet. While all bits of the packet are
successfully compressed without corruption, unlike the linear time delay approach
however, the loop based delay approach produces imperfect edges containing stray
bits in the compressed packet limited by the compression gating signal rise/fall times.
Additional encoding would therefore be necessary at the edge to prevent stray bits
from corrupting valid packet bits during decompression. Real time compression of
each packet may take place as read and write functions to the accumulative loop
buffer may be done simultaneously. It is therefore possible to achieve packet
compression with a high utilization at the compressor input.
114
time
Loop 2 output
Loop 3 output
Loop 4 output
SOA gated Output
Uncompressed Packet input
Compressed Packet output
time
Loop 2 output
Loop 3 output
Loop 4 output
SOA gated Output
Uncompressed Packet input
Compressed Packet output Figure 6. 15: Loop based implementation of fold-in compression (colors indicate virtual bins not
wavelengths)
6.4 Packet adaptation using fold-in compression
Due to the inherent advantages of fewer compression elements required, simpler
structure for scaling maximum input packet size and ease of construction, the loop
based approach is selected to demonstrate packet adaptation using the fold-in
compression approach. However, before setting up the compressor, characterization
of the SOA used to control and flush the compression buffer loop is necessary.
115
6.4.1 Loop characterization
The gated loop is accumulative by nature of the compression scheme and therefore
results in an increasing average power at the loop SOA’s input as shown in Figure 6.
16. The SOA must therefore be characterized to perform uniformly for a range of
input powers. For a maximum compression ratio of 1:4, the average input power for
the loop SOA varies over 6dB. The mark ratio of the input stream to the loop SOA
changes along with the average power. The SOA must therefore be characterized over
a sufficiently large pseudo random bit sequence.
time
Pav 2Pav 3Pav 4Pav
time
Pav 2Pav 3Pav 4Pav Figure 6. 16: Average Power increase at the input of the loop SOA
The loop SOA used in this demonstration was characterized using a 10Gbps bit error
rate system with a PRBS 231-1 sequence over a wide range of average input power.
Figure 6. 17 shows the power penalty incurred through the SOA for varying input
power. For lower optical powers, the power penalty incurred increases due to the low
input OSNR while at higher input optical power; pattern dependence of the SOA due
to saturation effects increases the incurred power penalty. It is therefore desired to
have an SOA with a high saturation power. A <0.5dB penalty is seen over the range
of -7dBm to -24dBm input power of ~17dB and is therefore sufficient to achieve a
maximum compression ratio of 1:4 with minimal degradation in SOA response over
the entire input packet. An average input power of ~-17dBm was chosen to operate
116
the compression loop. Due to the accumulative nature of the compression loop, a
higher compression ratio requires a wider input power range where the loop SOA
operates with minimal degradation. Compression from 2.5Gbps to 40 Gbps would
require a 12dB input power range for successful compression.
Operating region
Figure 6. 17: Measured SOA power penalty, OSNR Vs input power
6.4.2 2.5Gbps to 10Gbps Packet compression
The experimental setup to demonstrate 2.5Gbps to 10Gbps compression is shown in
Figure 6. 18. Compressed packet sizes fit a single sized bin of 1.2usec or the time
period of a 1500 byte packet at the compressed bit rate. The principle of operation of
the compression loop is as follows. 2.5Gbps packets to be compressed were generated
117
CW
Ring laser
10GHz clock source
BERT packet generator
Controller Card
Gain Modulated
SOA
Compression Ring
IsolatorASE Filter
Picosecond+ bulk Optical
Delay Line
50/50 coupler
Ring Gate
Ring Gate
Packet Gate
Compressor Gate
10GbpsCompressed
optical packet
2.5Gbps40-1500 byte
packet
To BERT gating
CW
Ring laser
10GHz clock source
BERT packet generator
Controller Card
Gain Modulated
SOA
Compression Ring
IsolatorASE Filter
Picosecond+ bulk Optical
Delay Line
50/50 coupler
Ring Gate
Ring Gate
Packet Gate
Compressor Gate
10GbpsCompressed
optical packet
2.5Gbps40-1500 byte
packet
To BERT gating
Figure 6. 18: 2.5Gbps to 10Gbps compression - Experimental setup
to be 1500 bytes (~4.8 usecs) in length with a 25 usec inter packet gap. The input
average power was set to an optimal level based on the characterization of the loop
SOA before entering the loop. A gain controlled SOA compensates for losses in the
loop and is used as a gating element to flush the compression loop when each packet
has been compressed. The SOA gating was adjusted to turn on the SOA only when
packets exist in the loop to eliminate noise buildup and amplifier transient response
due to bursty nature of the packet traffic. A 1.2nm filter centered at 1556.6nm was
used to reject ASE noise in the loop. A bulk fiber delay of ~210m and a picosecond
precision delay line were used to adjust the total loop length to be 1.2usec which is
selected as the bin size for this experiment. A variable attenuator was used to obtain
precise balance between gain and losses in the loop; this is critical to maintaining
118
constant bit heights between multiplexed bins. The compression loop serves as a
buffer to hold the packet while compression takes place as well as to time multiplex
the virtual bins of the packet as they enter the loop. A second gated SOA at the output
of the compression loop samples the compressed packet when all input virtual bins
have entered the loop and compression is complete. Picosecond control signals are
not required since time multiplexing is achieved by accurately controlling the length
of the compression loop. The gated SOAs used in this experiment have rise and fall
times of ~2 nsecs. Compressed packets generated in this experiment therefore have
2nsecs of stray bits at the packet edges.
Scope traces of the input packet and the compressed output packet are shown in
Figure 6. 19(a) and (b). The input packet size was measured to be 4.8usecs long and
the bit period to be ~400psec with a pulse width of ~17psec.
Figure 6. 19: (a) Input 1500 packet and (b) packet stream quality
119
Figure 6. 20: (a) Compressed 1500 byte packet and (b) compressed stream quality
Figure 6. 20(a) shows the 1500 byte packets after compression. The packet size was
measured to be ~1.2usec as expected from a 1:4 compression of the packet. The
compressed bit quality seen in Figure 6. 20(b) shows slight height variations in the
bits attributed to the loop SOA’s pattern dependence and a slight droop seen in the
loop control signal. Compressed bits are spaced 100psecs apart corresponding to a
compressed rate of 10Gbps.
Bit error rate measurements were made on the compressed 1500 byte packets at
10Gbps. The input packet stream bits were chosen such that after proper compression,
they would yield a 12000 bit PRBS 27-1 bit sequence. Error free operation was seen
for packets providing bit level verification of successful compression of the entire
packet. Bit error rate curves shown in Figure 6. 21 show the quality of the
compressed packets.
120
Figure 6. 21: Bit error rate measurements on 1500 byte compressed packets (refer Figure 6. 18)
A ~2.2dB penalty is seen through the compressor and is attributed to the height
variations of the compressed bits. This can be brought down by the appropriate choice
of an SOA with higher saturation power and better control signal quality.
6.4.3 Variable compression control
A second experiment was performed to test the compressor’s capability to handle
variable length input packets. Compression ratio may be varied by controlling the
number of loop trips being made by the incoming packet as explained in Figure 6. 22.
121
1500 byte or 12000 bit packet
Bit 0 Bit 12000
Incoming Packet at base rate
Compression loop Control for:
Pkt size< 375 bytes
Pkt size< 750 bytes
Pkt size< 1125 bytes
Pkt size< 1500 bytes
1500 byte or 12000 bit packet
Bit 0 Bit 12000
Incoming Packet at base rate
Compression loop Control for:
Pkt size< 375 bytes
Pkt size< 750 bytes
Pkt size< 1125 bytes
Pkt size< 1500 bytes Figure 6. 22: Principle of variable compression ratio control
By appropriate choice of the loop-gating signal, the compression ratio may be varied
to be 1:1, 1:2, 1:3 or 1:4. The compression scheme was tested for variable length to
fixed time slot adaptation of packets using variable compression ratio over an input
packet size range of 40 to 1500 bytes.
(a1)
(b1)
(a2)
(b2)
(a1)
(b1)
(a2)
(b2)
Figure 6. 23: 1500, 1024 byte input packets (a1,b1) and compressed output packets (a2,b2)
(1μsec time scale)
122
(a1)
(b1)
(a2)
(b2)
(a1)
(b1)
(a2)
(b2)
Figure 6. 23 and Figure 6. 24 show the input and compressed packet traces for 1500,
1024,560 and 40 byte packets. It can be seen that while the input packet size varies
from 4.8usecs (1500 bytes) to 128nsecs (40 bytes), a fixed output size of 1.2usec is
obtained after compression by varying the compression ratio from 1:4 to 1:1. For 40
byte packets, padding is required for the compressed packet to occupy a 1.2usec time
slot.
123
(c1)
(d1)
(c2)
(d2)
(c1)
(d1)
(c2)
(d2)
Figure 6. 24: 560, 40 byte input packets (c1,d1) and compressed output packets (c2,d2)
( 1μsec (top) and 50nsec (bottom) time scale
Bit error rates were not performed on smaller packets due to the minimum
synchronization time requirement of the bit error rate tester. Packet bit quality was
however verified visually to be the same or better than the 1500 byte compressed
packets.
6.4.4 Bandwidth efficient multiple compressor approach
The loop based compression approach has been demonstrated for 2.5Gbps to 10Gbps
compression of 1500 byte packets. Discretely variable compression ratio control on a
per packet basis required for variable length to fixed sized packet adaptation is
possible to achieve and has been demonstrated for individual packet sizes. However,
as studied in chapter 5, a more efficient bandwidth use at the core may be obtained by
124
compressing packets of different sizes to the maximum compression ratio. This
results in a discrete set of packet sizes after compression. The trimodal nature of the
distribution of packet sizes in today’s IP networks may be used in selecting
appropriate bin sizes to achieve maximum compression for most packets. When a
single delay line buffer is used, smaller compressed packets may be padded to the
maximum packet size at the compressed. However, in the case of buffers with
multiple delay line lengths, each compressed packet size may be buffered using the
delay line of optimal length thereby increasing the bandwidth efficiency. A two stage
wavelength converter architecture used in the core node shown in chapter 4 is
preferred to perform packet steering in the core forwarding due to its inherent
robustness and flexibility. Such an approach may also be useful in selecting the
output compressed packet size in the compressor. The polarization insensitive SOA
based XGM wavelength conversion may be used to select the appropriate internal
wavelength for transmission of the base rate packet. Bit carving is used to perform
NRZ to RZ conversion necessary before the packet can be compressed. A single loop-
based compressor may be used to implement compression to multiple compressed
packet sizes, based on input packet length as shown in Figure 6. 25.
125
Tunable λsource
Controller Card
Gain Modulated
SOA
Compression Ring
IsolatorASE Filter
Picosecond+ bulk Optical
Delay LineBased on λ
50/50 coupler
Ring Gate
Ring Gate
Compression select Compressor Gate
10GbpsCompressed
optical packetSOA
Tunable λsource
Controller Card
Gain Modulated
SOA
Compression Ring
IsolatorASE Filter
Picosecond+ bulk Optical
Delay LineBased on λ
50/50 coupler
Ring Gate
Ring Gate
Compression select Compressor Gate
10GbpsCompressed
optical packetSOA
Figure 6. 25: Bandwidth efficient compression setup.
Bins undergo a different delay based on the transmitted wavelength of the incoming
packet. Once the packet has been compressed, it may then be steered by the second
stage wavelength converter performing the actual forwarding. However, prior
knowledge of packet size is required at the control plane for the appropriate choice of
internal wavelength resulting in the optimal compressed packet size at the output.
This information may be transmitted in the optical header to ensure proper internal
wavelength selection before the packet arrives at the compressor input.
6.5 Chapter summary
The state of art in compression technology to date is presented in this chapter.
Deployment of compression in an optical network is considered and the requirements
126
for practical implementation of compression/decompression are studied. Prior
compression technologies based on the bit serial approach do not satisfy these
requirements. A new approach to compression, namely fold-in compression is
therefore proposed.
Two schemes are presented and compared to implement the fold-in compression
approach. While, the linear time delay approach offers clean packet edges for the
compressed packets, it does not scale well both with maximum achievable
compression ratio and maximum input packet length. The gated loop compression
scheme is shown to be a simple and robust design. It scales well with maximum bit
rate, highest compression ratio achievable and maximum input packet length.
Moreover, discretely variable compression ratio control required for variable length
to fixed sized packet adaptation is shown to be feasible in this scheme.
Compression of 1500 byte packets from 2.5Gbps to 10Gbps is shown with only a
~2.2dB penalty. Error free operation through the entire packet using the gated loop
approach was verified. Compression of variable length input packets to a fixed time
slot is shown by varying the compression ratio of the loop. A modification to the loop
based compression scheme is proposed to accommodate multiple compressed packet
sizes to increase the bandwidth efficiency of the forwarding scheme.
A compression technology capable of handling a maximum input packet length of
12000 bits (1500 bytes) and the widest dynamic range of input packet lengths from
320 bits (40 bytes) to 12000 bits (1500 bytes) shown to date is proposed and
127
demonstrated. It meets or exceeds all requirements for viable deployment of
compression as an adaptation mechanism at the core.
128
References
[1] K. Thompson, G. J. Miller, and R. Wilder, "Wide-area Internet traffic patterns
and characteristics," Network, IEEE, vol. 11, pp. 10-23, 1997.
[2] P. Toliver, D. Kung-Li, I. Glesk, and P. R. Prucnal, "Simultaneous optical
compression and decompression of 100-Gb/s OTDM packets using a single
bidirectional optical delay line lattice," Photonics Technology Letters, IEEE,
vol. 11, pp. 1183-1185, 1999.
[3] S. Aleksic, V. Krajinovic, and K. Bengi, "A novel scalable optical packet
compression/decompression scheme," presented at Optical Communication,
2001. ECOC '01. 27th European Conference on,478-479 vol.3, 2001.
[4] H. Sotobayashi, K. Kitayama, and W. Chujo, "40 Gbit/s photonic packet
compression and decompression by supercontinuum generation," Electronics
Letters, vol. 37, pp. 110-111, 2001.
[5] P. J. Almeida, P. Petropoulos, B. C. Thomsen, M. Ibsen, and D. J. Richardson,
"All-optical packet compression based on time-to-wavelength conversion,"
Photonics Technology Letters, IEEE, vol. 16, pp. 1688-1690, 2004.
[6] A. S. Acampora and S. I. A. Shah, "A packet compression/decompression
approach for very high speed optical networks," presented at
Telecommunications Symposium, 1990. ITS '90 Symposium Record.,
SBT/IEEE International 38-48, 1990.
129
[7] H. Toda, F. Nakada, M. Suzuki, and A. Hasegawa, "An optical packet
compressor based on a fiber delay loop," Photonics Technology Letters, IEEE,
vol. 12, pp. 708-710, 2000.
130
Chapter 7
Optical Packet Decompression
Compression of low speed packets into ultra-fast optical traffic is one approach to
tapping into the large bandwidth offered by optical networks. Such adaptation may
also be used to shape traffic characteristics such as packet size distribution to meet
today’s optical packet forwarding requirements. However, after forwarding, it is
impractical and sometimes unfeasible to use high-speed electronics to process
compressed packets directly before they exit the optical network. Decompression is
the process of re-expanding high speed packets to their original temporal sizes and to
base data rates that edge node electronics are capable of handling. In the previous
chapter, we proposed and demonstrated a compression scheme that is capable of
adapting variable length packets to fit into a fixed packet duration to better suit them
for optical buffer technologies available today. Such a compression scheme becomes
attractive only when a decompression mechanism that is scalable in packet size
handling capability is available after packet forwarding. This enables edge electronics
to process and transport packets error-free out of the optical network at base rate. In
this chapter, we identify the requirements of a packet decompression scheme for
successful deployment in an optical network and propose a technique capable of
operating in complement to the compression scheme discussed in chapter 6.
131
7.1 Decompression – State of the art
Access to the ultra-fast optical domain may be obtained by adapting low speed traffic
using packet compression and several approaches have been proposed and
demonstrated promising such adaptation. However, in order to successfully deploy
optical compression techniques and for low speed traffic (155Mbps to 2.5Gbps) to
benefit from the advantages of high-speed optics (10Gbps to 160Gbps and beyond), a
decompression mechanism is necessary to scale packets back to the bit rates handled
by electronics. While traffic entering the compressor is at the base rate, thereby
allowing for bit level manipulations with relative ease, packets into the decompressor
are at the compressed rate. At these ultra-fast data rates, complicated bit level
processing is impractical therefore setting a requirement for a higher level of
sophistication during decompression. Consequently, while several approaches have
been proposed for compression of packets, investigation into packet decompression
has been rather limited. As with the compressor, with the increase in the average
packet size and the packet size distribution of today’s internet [1], any optical
decompression scheme would have to be capable of handling different packet byte
sizes. In this section, we present the main approaches to decompression thus far and
investigate the limitations of prior methods in the context of a deployable scheme in
today’s optical packet network.
132
7.1.1 Feed-forward Delay line approach
In the delay line approach, high-speed compressed packets are injected into several
cascaded stages of delay lines to make multiple copies shifted in time. Appropriate
decompressed bits are then selected by gating at the compressed clock rate to yield
the output decompressed packet. Figure 7. 1(a) shows a typical setup for
decompression using this approach. [2] and [3] use this approach for packet
decompression and study the performance of such a decompression scheme. While
the implementation of the delay lines and the output gating element may differ, the
principle of operation behind both these schemes is the same as given in Figure 7.
1(b). Each delay line lattice element generates two copies of the input signal whose
…
(T-t) 2(T-t) 2N(T-t)Decompressed
packet OUT
Compressed packet
IN
- Coupler or switch element
High speed optical gating
…
(T-t) 2(T-t) 2N(T-t)Decompressed
packet OUT
Compressed packet
IN
- Coupler or switch element
High speed optical gating
(a)
1 2 3 4
21 43
21 43 21 43 21 43 21 43
Gating
Time aligned copies
Compressed input packet
Decompressed output packet
1 2 3 41 2 3 4
21 43
21 4321 43 21 4321 43 21 4321 43 21 4321 43
Gating
Time aligned copies
Compressed input packet
Decompressed output packet
(b)
Figure 7. 1: Feed forward delay decompression- (a) structure (b) operation
133
spacing is the geometrically increasing multiple of 2 times (T-t) where T is the base
rate bit spacing and t is the compressed rate bit spacing. The maximum number of bits
in each packet is determined by the number of cascaded delay line stages in the
decompressor and is given by N= 2M where M is the number of stages.
This maximum number of bits is however limited to (K-1) where K is the
compression ratio in the case of a simple decompression scheme. For packet sizes
greater than (K-1), use of parallel delay line structures would be necessary. However,
in this case the input compressed packet would have to be broken down to smaller
cells each with (K-1) bits, a process that would be extremely challenging at
compressed bit rates. Moreover, the number of delay lines cascaded would be limited
by the losses due to splitting and ASE noise accumulation if periodic or distributed
amplification was used to compensate for such losses. Though the implementation of
such a decompression scheme may be relatively simple for small packet sizes, this
approach does not scale well with packet size. Change in the compression ratio would
require a considerable modification to the setup and would therefore not be possible
on a per packet basis.
A variation of the delay line approach is presented in [4] where linear split and delay
of the input signal is used instead of the exponential approach. While integrated
electrical signal generator and optical gating elements are used, this approach also
does not scale well with packet size due to the linear nature of the decompression
scheme.
134
7.1.2 Spectral slicing
In this approach, the wavelength domain is used to create multiple time-delayed
copies of the compressed packets before time-gating the appropriate decompressed
bits as in [5]. Generation of multiple copies is achieved by creating a Supercontinuum
(SC). The SC is then spectrum sliced using an Arrayed Wave Guide Router (AWGR)
and time delayed such that each decompressed bit is aligned to the corresponding
base rate bit slot for every copy.
12 43 λ1 ... λ4
λin21 43
1 2 3 4λout
λ2 λ3 λ4λ112 43 12 43 12 43 12 43
12 4312 432 43 λ1 ... λ4λ1 ... λ4
λin21 43 λin21 43
1 2 3 4λout
1 2 3 4λout
λ2 λ3 λ4λ112 43 12 43 12 43 12 43
λ2 λ3 λ4λ112 4312 432 43 12 4312 432 43 12 4312 432 43 12 4312 432 43
Figure 7. 2: Decompression using spectral slicing
The output is then time gated using a saturable absorber and decompressed bits are
converted to the same wavelength using SC. The working principle for this approach
is shown in Figure 7. 2. The number of wavelengths used for decompression scales
linearly with the maximum number of bits in the packet and therefore severely limits
the packet size. Moreover, variable compression ratio would again require significant
setup changes and cannot be performed on a per packet basis.
135
7.1.3 Loop based approach
In this approach, a fiber loop is used to create multiple copies of the incoming
compressed packets as proposed in [6]. An electro-optic “AND” gate serves as the
gating mechanism to select the decompressed bit from each copy. Figure 7. 3 shows
the principle of operation of the loop based approach. The maximum packet size is
once again limited to (K-1) bits. ASE buildup in the loop can further limit the
decompressable packet size. In order to decompress larger packet sizes, a parallel
loop structure may be used. However, in this case, the complexity of the required
decompression control signals increases rapidly with packet size.
(T-t)
2 43
12 43 12 43 12 43 12 43 1 2 3 4
Gating
Decompressed packet
Compressed packet
(T-t)
2 432 43
12 43 12 43 12 43 12 4312 4312 432 43 12 4312 432 43 12 4312 432 43 12 4312 432 43 1 2 3 4
Gating
Decompressed packet
Compressed packet
Figure 7. 3: Loop based decompression structure and operation
A decompression mechanism chosen for deployment in an IP environment
would be required to handle packet sizes of up to 1500 bytes (12000 bits) and support
a wide dynamic range of packet byte sizes. Approaches proposed to date do not
handle more than 32 bits with little scope for scalability. The need for
minimization/elimination of complex ultra-fast control signals for decompression
suggests that we exploit the complementary features of compression/decompression
and select a compression scheme that simplifies decompression.
136
7.2 Fundamentals, design and implementations of packet
decompression
The choice of the approach to decompression of packets is primarily dictated by the
compression scheme. Depending on the nature of compression used, decompression
may be classified as decompression for fixed compression ratio, variable compression
ratio and discretely variable compression ratio packets. In chapter 6, we proposed and
demonstrated a discretely variable compression scheme to perform adaptation of
packets to higher speeds while compressing them to occupy the same fixed time slot
regardless of the input byte size. A complementing decompression scheme would
therefore handle packets occupying a fixed time slot and re-expand them to their base
rate packet size. Knowledge of the compression ratio used during compression is
therefore necessary for error-free decompression. In the intra-node architecture
network deployment of compression/decompression, such information is contained in
a single synchronized control plane that is aware of the location of the packet
throughout the core switching process. However, in the case of end-to-end
deployment, compression ratio information must be embedded in the packet header
for recovery before the packet arrives at the decompressor.
7.2.1 Decompression - principles of operation
Decompressed packets must contain all the bits of the original packet devoid of
corruption and in the correct bit order. Unlike compression, decompression involves
137
higher output utilization as compared to the input traffic due to the very nature of the
process as shown in Figure 7. 4.
Decompressor
Compressed packetsLow utilization at high bit rate
Decompressed packetsHigh utilization at low bit rate
Decompressor
Compressed packetsLow utilization at high bit rate
Decompressed packetsHigh utilization at low bit rate
Figure 7. 4 : Decompression principle - increase in utilization
In a packet environment, this sets a limit on the minimum interpacket spacing at the
input of the decompressor. The maximum input utilization for error free operation in
the case of fixed compression ratio input packet traffic is therefore given by
UIMAX = UO / K where K is the compression ratio.
In addition to this theoretical limitation, processing time for decompression may add
to the minimum required interpacket spacing. Packets entering the decompressor
prior to the minimum interpacket duration cannot be processed and a ‘timeout’
phenomenon occurs as in Figure 7. 5. Time out can therefore severely limit the packet
handling capacity of the decompressor. When the input utilization temporarily
exceeds UIMAX, buffering of incoming compressed packets may be used to resolve
timeout. However in the case when input utilization steadily exceeds UIMAX a
permanent solution such as the use parallel pipelining using multiple decompressors
feeding multiple base rate output ports would have to be considered to prevent packet
loss during decompression.
138
Time out issue
Compressed packets
Decompression
Time out issue
Compressed packets
Decompression
Figure 7. 5: Decompression - Time out issue
For discretely variable compression ratio implementation, interpacket spacing
requirements are determined on a per packet basis and given by
TSPACING = (K-1) * TPACKET
where TPACKET is the temporal size of the compressed packet. In the intra-node
deployment scheme, the control plane may use knowledge of the timeout requirement
of the decompressor to shape the traffic entering the decompressor after switching to
minimize packet loss.
7.2.2 Requirements on a packet decompressor
A decompression scheme capable of handling large packets and a wide dynamic
range of packet sizes should fulfill the following requirements:
• Decompression should require no complicated control signals at the
compressed bit rate. High speed control signaling would be limited to basic
sinusoidal waveforms or a combination of multiple sinusoids and preferably
driven by low voltage/current swing signals.
139
• Decompression processing time must be limited to reduce timeout overhead.
An ideal decompression scheme would process packets in real time thereby
supporting maximum line rate at the input.
• Scalability of the chosen scheme in terms of input compressed bit rate,
maximum packet size handling capability and input packet byte size dynamic
range is essential.
• The decompression scheme should not depend on elaborate packet and bit
level synchronization as this would negate the advantages of
compression/decompression.
• The number of control signals required must not scale directly with maximum
packet size or maximum compressed bit rate.
• Decompression of fixed cell size input packets with variable compression
ratios on a per packet basis would be required for error-free performance.
• Any approach to decompression must be robust to environmental effects for
successful deployment in an optical packet network.
7.3 Fold-out packet decompression
As with compression, implementing decompression may be approached using two
distinct techniques, the bit serial re-expanding approach and fold-out decompression.
Bit level manipulations that are required in the bit serial approach are practically
impossible to implement for larger packet byte sizes at high compressed bit rates due
to the lack of electronic control with the degree of precision and granularity required.
140
The fold-out technique requires few control signals and no complicated high speed bit
level manipulations and is therefore chosen for demonstration and study here.
Two separate implementations of fold-out packet decompression are proposed and
compared namely the linear time delay and the loop based decompression.
7.3.1 Linear time delay approach
In this approach, the wavelength domain is used to ease the requirement of numerous
electrical decompression control signals. A high speed Wavelength Converter (WC)
aids in separating virtual bins in the compressed packet. Modulating multiple
wavelength CW signals using a single compressed rate electrical clock with a packet
envelope signal generates an optical decompression signal. An AWGR and delay line
combination is then used to create the composite decompression gating signal at the
input of the WC. Compressed packets entering the decompressor are demultiplexed
and wavelength converted into multiple wavelengths as shown in Figure 7. 6. Spatial
separation of the demultiplexed bins takes advantage of the wavelength domain and is
done using an AWGR. Each bin is then appropriately delayed before being combined
into one bit stream. A wavelength insensitive RZ to NRZ converter is used to bring
down the decompressed packet bits to base rate pulse widths and rise/fall times before
being wavelength converted onto a single wavelength using an SOA based WC.
Decompressed packets are now ready to exit the node/network.
141
AWGR
Super--continuum or λ1,λ2,λ3,λ4 CW
Envelope delay
4 by 1 coupler
λ1λ2
λ3
λ4
Fiber WC
Compressed packet
Harmonics of base rate clock with Envelope of packet
AWGR
Decompression/ Fold-out delay
4 by 1 coupler
λ1- λ4 λ1λ2
λ3
λ4
RZ to NRZ
SOA-WCEgress
Base rateclock
CW @ λ0
AWGR
Super--continuum or λ1,λ2,λ3,λ4 CW
Envelope delay
4 by 1 coupler
λ1λ2
λ3
λ4
Fiber WC
Compressed packet
Harmonics of base rate clock with Envelope of packet
AWGR
Decompression/ Fold-out delay
4 by 1 coupler
λ1- λ4 λ1λ2
λ3
λ4
RZ to NRZ
SOA-WCEgress
Base rateclock
CW @ λ0
Figure 7. 6:Linear time delay decompression technique
The polarization dependence of the fiber WC imposes strict polarization control
requirements at the input of the decompressor to avoid penalties due to height
variation in the bits of different virtual bins within the decompressed packet.
Furthermore, while the technique enjoys the advantages of fold-out decompression,
the complexity of this implementation increases with number of bits in the
decompressed packet and the compression ratio, therefore making this approach less
scalable.
7.3.2 Loop based approach
We propose a second approach to implement fold-out decompression is the loop
based approach the working principle of which is detailed in Figure 7. 7. A
compressed packet entering the decompressor is stored in a packet buffer. A bit
shifted copy of the stored packet is output multiple times until the entire packet has
been decompressed. A decompressor gating is used to select the appropriate bits for
142
the virtual bins of the base rate packet to reconstruct the uncompressed packet at the
output. The requirement of a complex compressed bit rate control signal is eliminated
by bit shifting copies of the compressed packet and therefore requiring only a simple
repetitive gating signal at compressed rate and base repetition rate. Control of the
compressed packet buffer is achieved using base rate electrical control signals while
decompressor gating is driven using a periodic compressed rate electrical signal.
Bit shifted packet copier
Compressed packet buffer
Decompressor Gating
Compressed rate Base rate
Uncompressed packets
Compressed packets
Bit shifted packet copier
Compressed packet buffer
Decompressor Gating
Compressed rate Base rate
Uncompressed packets
Compressed packets
Figure 7. 7:Loop based decompression principle
Figure 7. 8 shows an example of packet decompression using the loop based
implementation to decompress a 1:4 compressed packet. A bit shifted copy is output
by the loop after each loop trip. At the decompressor gating unit, the bits
corresponding to each virtual bin of the decompressed packets align themselves to the
gating signal to result in decompression of the corresponding base rate bits of the bin.
After gating, the complete decompressed packet with the bits in the correct order is
recovered with little or no latency. By proper arrangement of the components in the
decompression loop, an output utilization close to 100% can be achieved for 1500
byte packet decompression. An RZ to NRZ converter may finally be used to widen
the decompressed packet bits to base rate widths before exiting the node/network.
143
Compressed Packet – fixedContainer size
…
Loop 2 output
Through the loop to buffer and time align packet for decompression
Loop 4 output
Cumulative Decompression loop output
Decompressed packet output
Bin gating using Fiber WC or LiNbO3 modulator
Decompression gating signal
Bit positions relative to the compressed packetCompressed Packet – fixedContainer size
…
Loop 2 output
Through the loop to buffer and time align packet for decompression
Loop 4 output
Cumulative Decompression loop output
Decompressed packet output
Bin gating using Fiber WC or LiNbO3 modulator
Decompression gating signal
Bit positions relative to the compressed packet
Figure 7. 8: Loop based fold-out decompression
Although the linear time delay approach affords the possibility of incorporating
wavelength conversion within the decompressor, the loop based approach was chosen
for experimental demonstration due to its scalability, ease of implementation and
compactness.
7.3.3 Loop characterization
Unlike the compression buffer loop, the decompression loop serves to only hold the
compressed packet as it is decompressed. The accumulative effect observed in the
compression loop in chapter 6 is therefore not seen here.
144
Figure 7. 9: Decompression SOA power penalty, OSNR, output power Vs input power
To pick the operating point of the SOA, a penalty measurement was done by varying
the average input power to the SOA as shown in Figure 7. 9. An operating point of ~-
18dBm is chosen for the decompression loop and the average input power into the
decompressor is appropriately adjusted. An average input power of -24dBm to -7dBm
or a range of 17dBm is seen for low distortion operation of the SOA (~0.5 dB p-p).
An operating point of ~-18dBm is chosen for the decompression loop to minimize
accumulating penalties within the loop and the average input power of packet traffic
into the decompressor is appropriately adjusted.
145
7.3.4 10Gbps to 2.5Gbps packet decompression
10G Ring Laser @ λ i
PC
Anritsu Pat Gen
SHF
Compressed Packets at 10G
ATTR
2x2
HP Pat Gen10 GHz Clk
ATTR
2.5Gb/s BER measurement
Gated SOA
BPF @ λ i
Bulk delay
Vardelay
ATTR
Iso.
DECOMPRESSION BUFFER DECOMPRESSION BUFFER LOOPLOOP
DECOMPRESSION DECOMPRESSION BIN GATINGBIN GATING
SHF
10G Ring Laser @ λ i
PC
Anritsu Pat Gen
SHF
Compressed Packets at 10G
ATTR
2x22x2
HP Pat Gen10 GHz Clk
ATTR
2.5Gb/s BER measurement2.5Gb/s BER measurement
Gated SOA
BPF @ λ i
Bulk delay
Vardelay
ATTR
Iso.
DECOMPRESSION BUFFER DECOMPRESSION BUFFER LOOPLOOP
DECOMPRESSION DECOMPRESSION BIN GATINGBIN GATING
SHF
Figure 7. 10: Loop based decompression - experimental setup
Decompression of 1500 byte packets from 10Gbps to 2.5Gbps was experimentally
demonstrated using the setup shown in Figure 7. 10. Compressed 1500 byte packets
at 1556.35nm were generated using a 10Gbps pattern generator and enter the
decompression buffer loop. An interpacket gap of 5usec is used between compressed
packets to ensure that they do not overlap after decompression. An SOA was used to
compensate for loop losses and is gated to eliminate transient gain behavior due to the
bursty traffic. An isolator and a 1.2nm band pass filter centered at 1556.35nm were
used to counter-propagation of signals and noise accumulation within the loop. A
bulk fiber delay of ~210m and picosecond variable delay were used to achieve the
buffer time. The loop delay was adjusted to be one compressed bit shorter than the
146
fixed time slot of the compressed packets in order to provide the appropriate bit shift
for the packet copies at the end of each loop trip. A Lithium Niobate modulator was
used to gate bits of each bin at the output of the decompressor to reconstruct the
original decompressed packet. Polarization sensitivity of the gating stage modulator
resulted in height variations between bits of separate bins.
(b1) (a1)
(a2) (b2)
Figure 7. 11: 10Gbps to 2.5 Gbps decompression of 1500 byte packets - (a1,b1) input compressed packets and bit quality, (a2,b2) decompressed packets and bit quality (2μsec (left) and 100psec
(right) time scale)
In the first experiment, 1500 byte packets at 10Gbps were input to the decompressor.
Figure 7. 11 (a1, b1) show the compressed packets occupying ~1.2usec and spaced
~5usec apart and the bit quality of the input packet spaced at 100psecs. Figure 7.
11(a2, b2) show the decompressed output packets occupying ~4.8usec and with a
~0.2usec interpacket gap. The decompressed bits can be seen to be of good quality
147
and spaced ~400psecs apart. Slight height variations seen in the decompressed
packets were due to the polarization sensitivity of the gating modulator and can be
reduced by using a polarization insensitive gating element. Bit error rate
measurements were made on the decompressed packets to verify the bit sequence
obtained and to assess the bit quality after decompression. The bit sequence in the
compressed 10Gbps 1500 byte packets was chosen to be such that after proper
decompression, a 12000 bit repeating PRBS 27-1 sequence would be formed that can
then be verified using a gated bit error rate detector.
Figure 7. 12 shows the bit error rate curves for the decompressed 1500 byte packets at
2.5Gbps. A receiver sensitivity measurement was performed for continuous 10Gbps
and 2.5Gbps traffic along with packet traffic at 2.5Gbps. A PRBS 231-1 sequence
used for sensitivity measurements shows no variation in receiver performance when
data rate was varied or when continuous or packet traffic was received. A penalty of
~2dB is seen after decompression of packets and can be attributed to the height
variations in bits of different virtual bins. It is important to note that after
decompression, a 96% output utilization is obtained for the 1500 byte packet stream,
indicating that the decompression scheme adds little or no overhead due to
processing. Decompression of packets performed on the fly is therefore verified.
148
Figure 7. 12: Bit error rate measurement on decompressed 1500 byte packets (refer Figure 7. 10)
7.3.5 Variable packet decompression
A second experiment to test the performance of the loop based decompressor was
made over the IP size distribution of 40 to 1500 bytes. Compressed input packets of
different sizes occupying a fixed time slot were input to the decompressor. The
compression ratio of the decompressor was varied by controlling the decompression
buffer signal. The decompressed packets (a2, b2, c2, d2) along with the compressed
inputs (a1, b1, c1, d1) for different packet lengths are shown in Figure 7. 13.
149
(a1) (a2)
(b1) (b2)
(c1) (c2)
(d1) (d2) Figure 7. 13: Compressed packet input(a1, b1, c1, d1) and decompressed packet output (a2, b2,
c2, d2) for 1500, 1024, 560 and 40 byte packets ( 500nsec time scale)
Interpacket spacing at the input is chosen to ensure that there is no packet overlap
after decompression. Decompression of 1500, 1024, 560 and 40 byte packets from a
fixed compressed packet size was successfully demonstrated using variable
150
decompression ratios of 4:1, 3:1, 2:1 and 1:1 respectively. Close to a 100%
decompressed link utilization was observed for most packet sizes. For packet sizes
occupying less than a bin, the utilization is limited only by the padding used at
compression. Bit quality for various packet sizes after decompression was visually
verified to be as good as or better than that of the decompressed 1500 byte packet.
7.4 Chapter summary
In this chapter, current approaches to packet decompression were examined and the
study shows that a decompression mechanism that is scalable in bit rate, maximum
packet size and range of packet size would require a novel approach with relaxed
requirements on the control signaling. We proposed and demonstrated a viable
scheme satisfying all the requirements for successful deployment of such a
decompressor in an IP based optical packet network. A performance study of the
decompression technique shows that packets up to 1500 bytes and a wide dynamic
range of packet sizes from 40 to 1500 bytes may be handled with a low power penalty
of ~2.0dB. on the fly processing of packets becomes critical to offset utilization
bottlenecks at the decompression point and we show that the loop based fold-out
decompression technique can operate at an almost 100% output utilization.
151
References
[1] K. Thompson, G. J. Miller, and R. Wilder, "Wide-area Internet traffic patterns
and characteristics," Network, IEEE, vol. 11, pp. 10-23, 1997.
[2] S. Aleksic, V. Krajinovic, and K. Bengi, "A novel scalable optical packet
compression/decompression scheme," presented at Optical Communication, 2001.
ECOC '01. 27th European Conference on,478-479 vol.3, 2001.
[3] P. Toliver, D. Kung-Li, I. Glesk, and P. R. Prucnal, "Simultaneous optical
compression and decompression of 100-Gb/s OTDM packets using a single
bidirectional optical delay line lattice," Photonics Technology Letters, IEEE, vol. 11,
pp. 1183-1185, 1999.
[4] H. Takenouchi, R. Takahashi, K. Takahata, T. Nakahara, and H. Suzuki, "40-
gb/s 32-bit optical packet compressor-decompressor based on an optoelectronic
memory," Photonics Technology Letters, IEEE, vol. 16, pp. 1751-1753, 2004.
[5] H. Sotobayashi, K. Kitayama, and W. Chujo, "40 Gbit/s photonic packet
compression and decompression by supercontinuum generation," Electronics Letters,
vol. 37, pp. 110-111, 2001.
[6] A. S. Acampora and S. I. A. Shah, "A packet compression/decompression
approach for very high speed optical networks," presented at Telecommunications
Symposium, 1990. ITS '90 Symposium Record., SBT/IEEE International 38-48,
1990.
152
Chapter 8
End-to-End Core Traffic
Adaptation
Compression and decompression of packets to achieve translation between bit rates
and as an adaptation mechanism from variable length to fixed time slots was
demonstrated in the previous chapters. However, in both cases, ideal traffic was
assumed at the input. A more complete understanding of the end-to-end performance
can be obtained by interfacing the output of the implemented compression stage to
the input of the decompression stage. Additional requirements to enable compression
and decompression may then be identified. In this chapter, we analyze and propose
additional adaptation requirements on packets for error free end-to-end compression
and decompression of variable sized packets. Finally, the performance of the end-to-
ed operation of compression and decompression is assessed both for fixed and
variable compression ratio implementations.
8.1 End-to-end core traffic adaptation
We have proposed and experimentally verified compression and decompression of
packets individually. However, in order to evaluate any issues in deploying this
153
technique to adapt core traffic, a complete end-to-end test of packets through the
compressor followed by the decompressor was performed. Packet distortion due to
pattern dependence during compression can be reduced by increasing the SOA
saturation power. The loop SOAs in the compression and decompression loops were
therefore swapped in order to optimize the end to end performance by taking
advantage of the higher saturation power of the decompression loop SOA. Timing
alignment of the two stages can also be improved by observing the output
decompressed eye diagram. When conducting experiments for compression and
decompression, ideal packets generated from a transmitter were used in both cases. It
was therefore not possible to analyze their behavior to an imperfect packet. Issues
relating to the end-to-end deployment of compression and decompression were first
analyzed.
8.2 Issues in end-to-end compression-decompression
experimentation
Signals entering the compressor were gated both by the loop SOA and the output
gating SOA. Any imperfections to the input packet, outside of the compression
window (packet duration) would therefore be suppressed. However, the base rate
rise/fall times of the control signals coupled with the control granularity achievable
result in an imperfect compressed packet at the output as detailed in Figure 8. 1. The
loop output present at all times during compression is eliminated by the gating SOA
at the output.
154
time
Loop 2 output Loop 3 output Loop 4 output
time
Uncompressed Packet input
Ideal gating signal –Compressed Rate precision
PRACTICAL gated Compression OutputPRACTICAL gated Compression Output
Typical gating signal –Control clock precision
Residual output
IDEAL gated Compression OutputIDEAL gated Compression Output
time
Loop 2 output Loop 3 output Loop 4 output
time
Uncompressed Packet input
Ideal gating signal –Compressed Rate precision
PRACTICAL gated Compression OutputPRACTICAL gated Compression Output
Typical gating signal –Control clock precision
Residual output
IDEAL gated Compression OutputIDEAL gated Compression Output
Figure 8. 1: Loop based fold-in compressor output imperfection
When an ideal gating signal with rise/fall times and control granularity smaller than
the compressed bit rate is applied, an ideal gating of only the compressed packet is
possible. A typical gating signal is limited by the control plane’s granularity and
rise/fall times along with the SOA’s gating drive circuitry. When such a signal is
applied to the output gate, residual outputs are passed through at the compressed
packet edges. When an ideal compressed packet with sharp edges enters the
decompressor, decompression without any bit corruption becomes possible. However,
when a compressed packet with residual outputs at the packet edges enters the
decompressor, bit corruption is possible at the virtual bin interfaces as shown in
Figure 8. 2.
155
Compressed Packet – fixedContainer size
Loop 2 output Loop 3 output Loop 4 output-- Corrupted Bit
Base Rate Gating
Loop output
Ideal CompressedSignal
Compressed Packet – fixedContainer size with residual outputs
Practical CompressedSignal
Loop output
Decompressed output
Decompressed output
Corrupted bits
Base Rate Gating
Compressed Packet – fixedContainer size
Loop 2 output Loop 3 output Loop 4 output-- Corrupted Bit
Base Rate Gating
Loop output
Ideal CompressedSignal
Compressed Packet – fixedContainer size with residual outputs
Practical CompressedSignal
Loop output
Decompressed output
Decompressed output
Corrupted bits
Base Rate Gating
Figure 8. 2: Loop based fold-out decompressor output imperfection
While higher precision of compression gating control reduces the number of bits
corrupted, it cannot eliminate corruption. To prevent loss of valid data, the optical
payload must be encoded at the network edge.
IP Packet
Payload (de)encapsulation
Compression (de)encoding
Optical packet (de)framing
IP Packet
Payload (de)encapsulation
Compression (de)encoding
Optical packet (de)framing
Figure 8. 3: (De)encoding to enable compression/decompression
The necessary encoding involves insertion of zero bits at the virtual bin interfaces as
shown in Figure 8. 3. Once compressed, the compression encoding provides a
156
sufficient guard band for the compression gating rise/fall times thereby eliminating
residual bits. Decompression without data corruption is therefore possible.
Loop 2 output Loop 3 output Loop 4 output
Typical gating signal –Control clock precision
PRACTICAL gated compression outputPRACTICAL gated compression outputwith edge compression coding with edge compression coding
Loop 2 output Loop 3 output Loop 4 output
Loop output
Base Rate Gating
Compression Coding – zeros at bin interfaces
Compression CodedInput Packet
PRACTICAL PRACTICAL decompression output decompression output
Loop 2 output Loop 3 output Loop 4 output
Typical gating signal –Control clock precision
PRACTICAL gated compression outputPRACTICAL gated compression outputwith edge compression coding with edge compression coding
Loop 2 output Loop 3 output Loop 4 output
Loop output
Base Rate Gating
Compression Coding – zeros at bin interfaces
Compression CodedInput Packet
PRACTICAL PRACTICAL decompression output decompression output
Figure 8. 4: Compression and decompression of encoded packets to prevent payload bit corruption
Encoded packets may then be compressed and decompressed without corruption of
payload bits as in Figure 8. 4 and adds to the packet overhead. By reducing increasing
the compression control precision, the core node encoding overhead may be brought
down. Decoding of the optical payload is necessary as it exits the network at the edge.
Payload encoding for core adaptation can be reduced to under ~2% by system
optimization.
A polarization insensitive Electro Absorption Modulator (EAM) was used in place of
the Lithium Niobate modulator to perform decompression gating to minimize height
variation in bits of the decompressed packets.
157
8.3 10Gbps core adaptation of 2.5Gbps traffic
The experimental setup used is shown in Figure 8. 5. The output of a fiber ring laser
with 13psec pulse width and at 1542.3nm was modulated by 2.5Gbps NRZ optical
packet data. The compression buffer loop length is selected to be one bin size
(~1.2µsec) offset by one compressed bit period to achieve proper virtual bin
interleaving. A Semiconductor Optical Amplifier (SOA) used to compensate for loop
losses, is gated to eliminate transient gain behavior due to the bursty traffic. The SOA
also serves to flush the compression loop when each packet has been compressed.
Fiber delays in the loop are chosen for the desired buffer time and a variable
attenuator is used to adjust the loop gain. A 1.2nm filter centered at 1542.3nm was
used to reject ASE noise in the loop. Interleaving of different sections (virtual bins) of
the incoming packet takes place at the output of the compression loop. A second
gated SOA is used to select the compressed packet with all bins completely
interleaved. The output of the compressor was then fed to a packet decompressor. The
decompressor buffer loop is used to make copies of the compressed packet that
occupy contiguous virtual bins. An EAM is used to gate appropriate set of bits at each
bin. However, the copy of the compressed packet occupying each bin must be shifted
in time by 1 compressed bit from the previous copy to align the correct set of base
rate bits in each virtual bin to the EAM gate.
158
Figure 8. 5: End -to-end compression/decompression experimental setup
BPF
FRL
BER
T Tx
Mod
10 G
HzPC
Attn
Cor
e C
ontr
olle
r
BPF Δλ
= 1
.2nm
Attn
Iso.
Com
p rin
g ga
teC
omp
outp
ut g
ate
Dec
omp
Rin
g G
ate
BER
T R
x Pa
cket
Gat
eC
omp
outp
ut
Gat
e
EAM
2.5G
b/s
BER
m
easu
rem
ent
BER
T R
x Pa
cket
Gat
e
N b
ytes
N b
ytes
N b
ytes
N b
ytes
N b
ytes
N b
ytes
2.5G
bps
Pack
ets
10G
bps
Pack
ets
2.5G
bps
Pack
ets
SOA
2x2
Attn Var.
DelayBu
lk
Dela
y
BPF Δλ
= 1
.2nm
Iso.D
ecom
p R
ing
Gat
eD
ecom
pres
sion
B
uffe
r Loo
pEDFA
2x2
Attn Var.
Delay
Bulk
De
lay
Iso.C
omp
Rin
g G
ate
SOA
Com
pres
sion
B
uffe
r Loo
p
BPF Δλ
= 1
.2nm
Dec
omp
bin
Gat
ing
BPF
FRL
BER
T Tx
Mod
10 G
HzPC
Attn
FRL
FRL
BER
T Tx
BER
T Tx
Mod
10 G
HzPC
Attn
Cor
e C
ontr
olle
rC
ore
Con
trol
ler
BPF Δλ
= 1
.2nm
Attn
Iso.
BPF Δλ
= 1
.2nm
Attn
Iso.
BPF Δλ
= 1
.2nm
Attn
Iso.
BPF Δλ
= 1
.2nm
Attn
Iso.
Com
p rin
g ga
teC
omp
outp
ut g
ate
Dec
omp
Rin
g G
ate
BER
T R
x Pa
cket
Gat
eC
omp
outp
ut
Gat
e
EAM
EAM
2.5G
b/s
BER
m
easu
rem
ent
2.5G
b/s
BER
m
easu
rem
ent
BER
T R
x Pa
cket
Gat
e
N b
ytes
N b
ytes
N b
ytes
N b
ytes
N b
ytes
N b
ytes
N b
ytes
N b
ytes
2.5G
bps
Pack
ets
10G
bps
Pack
ets
2.5G
bps
Pack
ets
N b
ytes
N b
ytes
N b
ytes
N b
ytes
2.5G
bps
Pack
ets
10G
bps
Pack
ets
2.5G
bps
Pack
ets
SOA
Iso.D
ecom
p R
ing
Gat
e
2x2
Attn Var.
DelayBu
lk
Dela
y
BPF Δλ
= 1
.2nm
Dec
ompr
essi
on
Buf
fer L
oop
SOA
Iso.D
ecom
p R
ing
Gat
e
BPF Δλ
= 1
.2nm
2x2
2x2
Attn Var.
DelayVar.
DelayBu
lk
Dela
y
Bulk
De
layDec
ompr
essi
on
Buf
fer L
oopED
FA2x
2
Attn Var.
Delay
Bulk
De
lay
Iso.C
omp
Rin
g G
ate
SOA
Com
pres
sion
B
uffe
r Loo
p
BPF Δλ
= 1
.2nm
2x2
Attn Var.
Delay
Bulk
De
lay
Iso.C
omp
Rin
g G
ate
SOA
Com
pres
sion
B
uffe
r Loo
p
BPF Δλ
= 1
.2nm
2x2
2x2
Attn Var.
DelayVar.
Delay
Bulk
De
lay
Bulk
De
lay
Iso.C
omp
Rin
g G
ate
SOA
Com
pres
sion
B
uffe
r Loo
p
BPF Δλ
= 1
.2nm
Dec
omp
bin
Gat
ing
159
This is achieved by adjusting the buffer loop size to be one bin size offset by one
compressed bit period. Figure 8. 6(a1, a2) show incoming 1500 byte packets at
2.5Gbps occupying 4.8µsec with bits spaced 400psecs apart and being compressed to
10Gbps with 100psec bit spacing and occupying a fixed time slot of ~1.2µsec shown
in Figure 8. 6(b1, b2). Figure 8. 7(b1, b2) shows the decompressed packet occupying
4.8usecs with a bit spacing of 400psec.
(a1) (a2)
(b1) (b2) Figure 8. 6: 1500 bytes (a1, a2) Input packet ,bit quality at 2.5Gbps (b1, b2) compressed packet, bit
quality at 10Gbps
Both compression and decompression schemes are now scalable to higher
compressed bit rates and compression ratios due to the simple bin alignment nature of
the compression and decompression schemes and their high tolerance to slow edges
and granularity of control signals.
160
Figure 8. 7: 1500 byte packets (a1, a2) compressed packet and bit quality at 10Gbps, (b1, b2)
decompressed packet and bit quality at 2.5Gbps
To measure end-to-end compressor/decompressor BER performance, a 1500 byte
packet stream containing PRBS 27-1 payload was used. BER measurements were
made for packets constructed from a PRBS 27-1 sequence for proper sequence
detection for integrity check. 1500 byte packets were used due to the minimum
synchronization time requirement of the BERT receiver.
The received decompressed packets showed error free performance when sent to a
gated bit error receiver, confirming the integrity of the received packet. A receiver
sensitivity measurement was made on 2.5Gbps packets with a PRBS 27-1 sequence
generated directly from the transmitter and was found to be ~-18.6dBm. A ~1.65dB
power penalty is seen in the BER measurements shown in Figure 8. 8 from the test on
packets undergoing both the compression and decompression. SOA pattern
161
dependence and bin amplitude variations were identified to be caused of penalty in
the measurement.
Figure 8. 8: End-to-End bit error rate measurement at 2.5Gbps
8.4 Variable length packet adaptation
Variable compression ratios (1:1, 1:2, 1:3 and 1:4) necessary to implement
compression and decompression of variable byte size packets to and from a fixed
compressed packet time slot may be achieved by controlling the length of the ring
gating signals of both the compressor and decompressor. End-to-end performance
was tested for variable input packet sizes from 40 to 1500 bytes.
162
Input Compressed Decompressed
(a)
(b)
(c)
(d)
Figure 8. 9: Input, compressed and decompressed packets for (a) 1500 byte, (b) 1024 byte, (c) 560 byte, (d) 40 byte packets
Figure 8. 9 shows the input, compressed and decompressed packets for (a) 1500 byte,
(b) 1024 byte, (c) 560 byte and (d) 40 byte packets. Compressed packets occupy a
1.2usec time slot regardless of input packet size. It can be seen that the decompressed
163
packets are the same size of the original base rate packets as is expected. 40 byte
packets are padded to occupy the 1.2usec time slot.
8.5 Chapter summary
End to end deployment of compression and decompression techniques are studied and
we identify an adaptation requirement to prevent data corruption due to the use of
control signals with slow rise/fall edges. We demonstrate that compression and
subsequent decompression of packets from 2.5Gbps to 10Gbps and back to 2.5Gbps
may be performed for the entire IP packet byte size range commonly found in today’s
traffic with a low end to end penalty of ~1.65dB. Both techniques are scalable to
higher bit rates and packet sizes while still being controlled by base rate electrical
signals. Adaptation to and from fixed cell sizes suitable for optical buffering
technologies for a wide range of packet byte sizes has been shown to be possible with
a low penalty and with little or no processing time overhead.
164
Chapter 9
Conclusion
9.1 Thesis conclusions
With the increasing capacity demands placed on today’s internet routers, more
functionalities are being moved from the electronic to the optical domain. While
optical integration of these forwarding functionalities has been a promising area of
research, the need for research in adaptation to enable forwarding of IP packets using
optical technologies is evident from the fact that no optical forwarding of IP traffic
has been demonstrated to date. Consequently, a majority of the research presented in
this thesis has been devoted to investigating the issues involving such adaptation that
is critical to the practical realization of end-to-end-IP forwarding through the use of
all-optical technologies.
Several state of the art approaches to optical transport and forwarding were evaluated
from the perspective of identifying the requirements in building a mechanism that
would support IP traffic while packet switching is performed at the core. When
building such an optical network, we realize that performance may no longer be
evaluated only on the basis of physical layer metrics such as Optical Signal to Noise
Ratio and Bit Error Rates, but will have to include network level testing such as
165
packet throughput and latency measurements. In order to perform forwarding at the
optical core, a framing mechanism must achieve complete decoupling of the optical
payload from the header and framing information. A framing structure is proposed
that started based on requirements identified from previous work but has gone
through several stages of evolution from experimental work with IP over optics. It
takes into account various signal processing uncertainties and bandwidth
considerations and was designed with payload bit rate transparency in mind. Ingress
and egress nodes require several optical and electronic sub-systems to implement the
proposed framing mechanism and we researched the influence on them by the
physical signaling characteristics of the packet steam. The need for idlers is shown to
be extremely critical both at the core and at the network edges for proper functioning
of both optics and electronics alike. Successful implementation of the edge nodes
(ingress and egress) was shown and the back to back demonstration of edge node
adaptation was presented. Throughput measurements on traffic through the
implemented edge adaptation show that the experimental results closely match
theoretical predictions for all packet sizes. A minimum throughput of ~40% and a
maximum throughput of ~95% are observed demonstrating the feasibility of such
adaptation without major penalties in packet forwarding rates. We also observe that
latency added at the edge is significant but varies in a fairly linear fashion with
increasing packet size. We note that this changes when a more complex lookup
mechanism is used. We identify switching times and gain control times are major
influencing factors in the optical domain in determining the packet overhead and
166
consequently the forwarding throughput of the core. Electronic bottlenecks include
control signal rise/fall times and control granularities based on temporal uncertainties
and observe that guard band sizes are affected significantly due to the lack of
appropriate electronic technologies. From research in optical packet forwarding
implementation, we see the need for specialized electronic modulator and current
drivers that can support varying signal dc content even when idlers are used to
stabilize packet streams. This may be possible by either dc-coupling high speed
electronics or reducing the low frequency cut off’s of the existing ac coupled
circuitry. Throughput measurements on IP streams show that minimal penalty is
imposed by the forwarding process in the core and end-to-end performance match
theoretical predictions very well. Latency measurements were performed through the
core forwarding node and compared with the back-to-back edge node latency
measurements and we observe that only a 0.79usec additional latency is introduced at
the core as compared to the 10 to 40usec latency through the edge. It can therefore be
deduced that the major contribution to end-to-end IP latency arises from the edge and
cascading several core nodes would result in minimal impact.
While an edge adaptation and a core node forwarding mechanism that supports
variable sized IP packets was demonstrated, we observe the need for fixed time slot
payloads in order to efficiently buffer packets at the core to resolve contention.
Compression and decompression was proposed as a means to speed up and adapt
variable byte seized packets into fixed time slots and an intra node deployment
strategy is suggested. We propose a new approach to packet compression namely the
167
fold-in compression that is capable of handle IP packet sizes of up to 1500 bytes and
more. Compression of variable sized packets from 40 to 1500 bytes from 2.5Gbps to
10Gbps is shown using the loop based implementation of the fold-in approach and we
observe a 2.2dB power penalty due to compression. This is the largest packet size and
the widest range of input packet lengths compressed to date. We also propose and
demonstrate a decompression scheme that complements the fold-in compression
approach, capable of decompressing 40 to 1500 byte packets from a fixed time slot at
10Gbps to their original packet sizes at a base rate of 2.5Gbps. A ~2dB power penalty
is observed through the decompressor for 1500 byte packets. Finally, we identify the
issues associated with implementing end-to-end packet compression and
decompression and propose a modification to the edge node framing to support
compression/decompression at the core. We experimentally demonstrate the
performance of back-to-back compression and decompression for 40 to 1500 byte
packets and reduce the overall power penalty to ~1.65dB for 1500 byte packets by the
use of a polarization independent switching stage at the decompressor output.
Feasibility of using compression and decompression both to speed-up the packet
payload and thereby reducing its temporal footprint and to optically adapt variable
byte sized packets seen in today’s IP networks (40 to 1500 bytes) into a fixed time
slot. The proposed compression and decompression schemes are scalable both in the
maximum input packet byte size and in the highest achievable compressed bit rate
and can therefore be used for scaling payloads bit rates to 40 Gbps or higher and may
be modified to support bigger packets sizes such as jumbo packets of 9000 bytes.
168
9.2 Future work
Several ways to increase the overall optical packet throughput may be taken such as
increasing the interface bit rates to 10Gbps. While the payload and framing data rates
implemented in the research presented in this thesis were the same, the edge
adaptation mechanism may be tested for its performance when frame and payload
data rates are different (e.g. 10Gbps header and 40Gbps payload). Timing uncertainty
due to electrical serialization and deserialization processes can be reduced
significantly by working with electronic chip vendors to meet the additional needs to
gain tighter control over the output signal timing. It could be interesting to investigate
the possibility of reducing overhead due to optical rise/fall times and the subsequent
impact on throughput performance.
While compression and decompression has been proposed and demonstrated to
10Gbps, theoretical and experimental performance analysis of scaling the payload
speed up to 40Gbps and the compression ratio to 1:16 or higher could prove
beneficial extremely useful. Moreover, NRZ-to-RZ conversion and RZ-to-NRZ
conversion required to enable compression and decompression was not demonstrated.
The scheme as proposed would be more suited for edge node deployment but can be
modified suitably using clock recovery for core node deployment. NRZ-to-RZ
conversion may be done electro-optically as suggested in chapter 5 but an all-optical
approach would make core node deployment of compression and decompression
simpler and could be investigated. A demonstration of a multiple-input multiple-
output core node would take OPS technology closer to practical realization. While
169
optical adaptation has been demonstrated, implementing buffering using recirculating
fiber delay loops using adapted packet traffic and a comparison of its packet
throughput performance against packet forwarding with no contention resolution
mechanism would be interesting.
170