Exfo Guide 100g-400g en
Exfo Guide 100g-400g en
Exfo Guide 100g-400g en
About EXFO
EXFO develops smarter network test, monitoring and analytics solutions for the world’s leading communications
service providers, network equipment manufacturers and webscale companies. Since 1985, we’ve worked
side by side with our customers in the lab, field, data center, boardroom and beyond to pioneer essential
technology and methods for each phase of the network lifecycle. Our portfolio of test orchestration and real-
time 3D analytics solutions turn complex into simple and deliver business-critical insights from the network,
service and subscriber dimensions. Most importantly, we help our customers flourish in a rapidly transforming
industry where “good enough” testing, monitoring and analytics just aren’t good enough anymore—they never
were for us, anyway. For more information, visit EXFO.com and follow us on the EXFO Blog.
Table of Contents
1. High‑Speed Technology Overview............................................................................................................................ 2
1.1. Introduction to 100G Technology and Market Drivers................................................................................................... 2
1.2. 40G/100G Ethernet (IEEE 802.3ba)................................................................................................................................. 3
1.3. Optical Transport Network/ITU‑T G.709............................................................................................................................. 4
1.4. 40G/100G Pluggable Optics................................................................................................................................................ 5
1.5. CLR4/CWDM4 100G Transceiver Specifications.............................................................................................................. 7
1.6. 400G Ethernet (IEEE 802.3bs).......................................................................................................................................... 8
1.7. 400G Physical Interfaces...................................................................................................................................................11
1.8. Flex Ethernet......................................................................................................................................................................... 12
2. 40G/100G System Testing and Lab Qualifications................................................................................... 16
Line Side
2.1. Transmitter Compliance Verification ............................................................................................................................... 16
2.2. Receiver Performance Verification...................................................................................................................................18
Client Side
2.3. Physical Layer Test and Transceiver Testing.................................................................................................................20
2.4. OTU Frame Structure, Overhead and Testing................................................................................................................. 34
2.5. Forward Error Correction (FEC)........................................................................................................................................ 38
2.6. Response Test, OTN Equipment Conformance and Interoperability Test................................................................39
2.7. ODU‑MUX Testing: Multiplex Mapping Test Method for Constant Bit‑Rate Services (e.g., SDH).......................... 41
2.8. Packet/Ethernet over OTN Test: ODU0, GMP ODUflex and Hitless Adjustment of ODUflex (GFP) (HAO)........... 42
2.9. 40G/100G Ethernet Services over OTN Testing.............................................................................................................49
2.10. 40G/100G OTN Service‑Disruption‑Time Measurement ............................................................................................ 50
2.11. Delay Measurement............................................................................................................................................................53
2.12. TCM and Performance Monitoring....................................................................................................................................54
2.13. Advanced Intrusive Through‑Mode Analysis.................................................................................................................55
2.14. OTN Performance Testing: Long‑Term Bit‑Error Test.....................................................................................................56
2.15. Multichannel OTN.................................................................................................................................................................58
2.16. General Communication Channels BERT Test................................................................................................................64
2.17. Ethernet Testing in the Lab: RFC 2544 Test...................................................................................................................65
2.18. Ethernet Service Aggregation and Switching Capability of POTN Devices............................................................... 67
2.19. Multilink................................................................................................................................................................................69
3. 40G/100G Commissioning and Turn‑Up............................................................................................................ 71
Line Side
3.1. Link Characterization and Non‑Linear‑Effect Validation.............................................................................................. 71
3.2. OSNR Measurements and Polarization‑Dependent OSNR Verification..................................................................... 80
3.3. BER Test Using Receiver‑Provided Information.............................................................................................................89
Client Side
3.4. Inspecting Fibers and Connectors ................................................................................................................................. 90
3.5. Fault Sectionalization and CFP Health Check.................................................................................................................92
3.6. Continuity Testing............................................................................................................................................................... 95
3.7. OTN Long‑Term BERT........................................................................................................................................................... 95
3.8. ITU‑T Y.1564 for Ethernet Services Testing.................................................................................................................... 96
3.9. 40GE/100GE Troubleshooting Using Filtering and Packet Capture ....................................................................... 106
3.10. Ethernet Service OAM (SOAM) ....................................................................................................................................... 109
3.11. POTN and MPLS‑TP Test................................................................................................................................................... 116
3.12. POTN Packet/Ethernet Service Quality of Service Verification................................................................................ 118
3.13. Clock Performance Test for Packet‑Switching‑Based POTN Device in Response to ODUk/ODUflex Client Signal ...... 119
3.14. Multiservice Application Characterization in OTN and POTN Networks.................................................................. 121
3.15. Ping and Traceroute from 40GE/100GE Interface...................................................................................................... 122
4. Acronyms—Transport and Datacom..................................................................................................................123
One aspect that the OIF took into consideration is that these new 100G systems should be
able to share the physical medium with existing 10G and 40G signals. In addition, several of
the networks already had multiple reconfigurable add‑drop multiplexers (ROADMs) and 50 GHz
channel spacing and therefore, the new 100G signals would need to be compatible with this
architecture. For the most part, these links had been compensated for dispersion, and this factor
was also taken into account. All of these considerations contributed to a very precise modulation
technique.
All other modulation schemes were quickly eliminated, and the OIF advised the industry to use
dual polarization or Pol‑Mux QPSK, also known as NRZ‑DP‑QPSK, DP‑QPSK or PM‑QPSK,
as the modulation of choice for 100G transmission. Consequently, transmission started moving
toward a combination of amplitude and phase modulation, contributing to yet another new factor
to be taken into account within the context of testing.
In the process of transmitting a 40G/100G Ethernet signal (see Figure 2), using a 100GBASE‑LR4
CFP, the Ethernet packet is broken into blocks that are mapped into PCS lanes using a
round‑robin distribution. To ensure a clean realignment of the entire block while managing the
skew between blocks at the receiving end, a marker block will be added on each logical lane; this
marker is sent at a fixed cycle of 210 µs. The twenty PCS logical lanes will then be multiplexed
down to ten lanes at 10 Gbit/s (also called CAUI lanes) using the physical medium attachment
(PMA) located in the C form‑factor pluggable (CFP) optical interface. The unique property of
the PCS lanes is that regardless of how they are multiplexed, all bits from the same PCS lane
follow the same physical path. This enables the receiver to correctly reassemble the aggregate
channel. At the same time, the unique lane marker enables the deskew operation in the receiver.
The 40G/100G Ethernet implementation introduces new challenges at both the electrical and
optical layers. These challenges are due to the data distribution over multiple channels combined
with the skew effect. This triggers the need to test the individual concepts, such as PCS lanes,
PCS skew and alignment markers, to ensure that the 40G/100G Ethernet device under test
supports the proper realignment capabilities, thus compensating for the physical characteristics of
the link. By inserting multiple unit intervals (UIs) of skew on different lanes, engineers can identify
receiver buffer issues, which is crucial when validating the skew tolerance of the 100G system.
2:1
2:1
2:1
2:1
2:1
2:1
2:1
2:1
2:1
Similar to the layers defined by the IEEE P802.3ba Task Force, the ITU‑T SG15 defined a new
layer on top of the optical channel transport unit (OTU) layer, known as the optical channel
transport lane (OTL) protocol. The OTL layer is defined as OTL x.y, where “x” represents the data
rate and “y” represents the number of optical lanes. For example, OTL 4.4 represents an OTU4
signal running on four wavelengths at 28G. OTL introduces elements, such as OTL lanes, OTL
skew and OTL alignment markers, which are conceptually similar to the PCS lanes, PCS skew
and PCS alignment marker described in the IEEE P802.3ba standard. As with Ethernet, the OTL
layer uses logical lane markers to realign all sent blocks: the main differences in this case is that
the marker is embedded in the frame alignment signal (FAS) and multiframe alignment signal
(MFAS) bytes of the OTN frame. Therefore, when validating 100G devices supporting OTN
capabilities, it is crucial to test the skew at the OTL layer, in addition to testing all other OTN
layers (including OTU4, ODU4, OPU4 and FEC), to ensure proper mapping and demapping of
100G Ethernet client signals and reporting of fault management.
λ1 λ2 λn
Qualifying the performance of the OTN physical layer by testing the line rate of the signal with
a pseudo random bit sequence (PRBS) pattern is a key step. This test is critical for designers
and system engineers during the development of the 100G line cards. One of the main tests
involves injecting a complex PRBS pattern on each of the physical lanes and analyzing the bit
error rate (BER) at the receiver end. The ability to add a different pattern per lane provides a
whole overview of the crosstalk between the channels. Moreover, a maximum density pattern can
be configured to verify the phase‑locked loop (PLL) receiver’s response to unbalance bit density.
Technology has evolved and allowed the introduction of newer 100G transceivers that consume
less power and increase port density at a lower cost, namely CFP2, CFP4 and QSFP28. Unlike
the first CFP transceiver, CFP2, CFP4 and QSFP28 are not equipped with the gearbox module,
which is now present on the host equipment (i.e., network element equipment or test equipment).
Each one of these new transceivers receive 4 CAUI‑4 lanes in the electrical interface, which are
directly converted into 4x25G optical lanes. These transceivers are all interoperable as long as
their optical characteristics are compatible: SMF versus MMF, wavelength, number of lambdas
and power level sensitivity. It’s worth mentioning that for long reach applications, the most
commonly used optical interface is the LC connector; for short reach applications, the most
commonly used interface is the MPO (or MPT) connector.
PCS/Logical CAUI/Physical
Lanes Lanes PMD Lanes
2:1
2:1
2:1
2:1
2:1
2:1
2:1
2:1
LAN
Electrical/
2:1
2:1
10:4 Optical
WDM
(optical mux)
2:1
2:1
PCS/Logical MPO-12
Lanes
2:1
2:1 CAUI-
4 Lanes MPO connector
2:1
2:1
2:1
2:1
10:4 Electrical/
Optical MMF Fiber
2:1
2:1
2:1
2:1
CFP2, CFP4 and QSFP28 transceivers are also supported by the CFP Health Check application
and their qualification and testing remain a very significant task when deploying 100G links.
CLR4
100G CLR4 is a multivendor specification for 100 Gbit/s (4 x 25.78 GBd) focusing on providing
solutions to fill data center requirements. It is a low‑power, cost‑effective solution for transmission
distances up to 2 km using duplexed single mode fiber.
CLR4 transceivers are able to operate with and without forward error correction (FEC).
Functioning with FEC ensures interoperation with CWDM4 MSA.
This specification is transceiver form‑factor agnostic, however the introductory form factor is a
QSFP and is based on a CWDM grid with 20 nm spacing to allow lower temperature operation
(as defined by IEEE 802.3 CL87 40GBASE‑LR4).
More and more data centers are being built to support high‑capacity data networks. These
centers are interconnected using 100 Gbit/s links but will soon migrate to 400 Gbit/s to increase
capacity. According to the high‑speed roadmap published by the Ethernet Alliance (Figure 9),
400G Ethernet (400GE) is poised to become the next client rate in the Ethernet ecosystem
as the industry ramps up to handle the massive demands of hyperscale data centers, service
providers and business users.
IEEE is developing a new standard for 400GE (IEEE 802.3bs) offering improved capabilities for
key aggregation and high‑bandwidth interconnect applications, including cloud‑scale data centers,
internet exchanges, colocation services, wireless infrastructure, service provider and operator
networks and video distribution infrastructure.
This application note will highlight the key objectives of the new 400GE rate and the differences
relative to 100GE. We will also take a look at the new supported physical interfaces.
The 400G PCS lane skew, where we measure the difference between the times of the earliest
PCS lane and latest PCS lane for the transition of the alignment marker sync bits.
Threshold crossing
FEC excessive threshold
• If the pre‑FEC BER exceeds this mandatory threshold for longer than the Interval, link fault
signaling is generated
L Fault Det −
L Fault Rcd −
Remote Fault −
LOA −
Hi‑SER −
L Deg SER Det −
L Deg SER Rcd −
R Deg SER −
• 500 m SMF: 200GBASE‑DR4: 200 Gbit/s transmission over four lanes (i.e., 8 fibers total)
• 2 km SMF: 200GBASE‑FR4: 200 Gbit/s transmission over a 4 WDM lane
(i.e., 2 fibers total)
• 10 km SMF: 200GBASE‑LR4: 200 Gbit/s transmission over a 4 WDM lane
(i.e., 2 fibers total)
An objective of this new implementation agreement is to improve and maximize the interconnection
between network elements (routers) and transport gear. A second objective is to help network
elements reach the bandwidth being currently handled today by transport elements that use
coherent technologies.
Let’s look at the elements that allow this flexible physical mapping structure to work:
FlexE client: A FlexE client is defined as an Ethernet data stream associated to a rate that could
(or not) belong to an existing Ethernet PHY rate (see Figure 15). FlexE clients defined by the
agreement include Ethernet rates already standardized (10G, 40G, 100G) as well as new rates
(e.g., 200G and 400G) and finally multiples of 25G(n x 25G) such as 125 Gbit/s etc. Each FlexE
client has a separate MAC address.
FlexE group: The OIF defines the FlexE group as the “n” Ethernet PHYs bonded on the link
where each group should have at least one Ethernet PHY. The first version of the implementation
agreement defined 100G PHYs, but upcoming versions may include additional PHYs based on
new standard rates.
The PHYs belonging to the FlexE group must have a number that identifies them within the group
and also must be interconnected by the same two FlexE shims.
FlexE shim: Defined as the layer located between the MAC and the PCS inside the IEEE 802.3
stack (see Figure 16), FlexE shim is responsible for associating the different clients transported
over a FlexE group.
The clients are distributed sequentially into the calendar inside the FlexE shim and then distributed
to each physical PHY PCS lane where each PHY is transported independently over the network.
The FlexE shim is responsible for calendar management, synchronization and PHY mapping.
As part of the implementation agreement, the illustration below shows the main actions that a
network element can do with FlexE:
Bonding: Bonding creates high capacity data pipes by assigning several Ethernet PHYs to
transport MAC rates greater than 100G. (see Figure 17)
Sub‑rating: An action that allows network elements to sub‑divide physical interfaces in order
to transport lower data pipes over partially filled Ethernet PHYs. (see Figure 18)
Compliant with the OIF FlexE implementation agreement, this latest Power Blazer module includes
four QSFP28 ports and is a robust testing solution for today’s challenges, with capabilities that
include:
LINE SIDE
The next section describes the first stage of testing, which primarily involves R&D and
manufacturing, and also outlines how manufacturers intend to test both transmitters and receivers
once the volume starts to increase.
A QPSK modulator is based on two Mach‑Zehnder modulators, as illustrated in the middle of the
block diagram on the next page. These modulators both use the same carrier‑wave laser, and the
bottom branch is phase‑shifted by 90 degrees to generate the “I” (in‑phase) and “Q” (quadrature)
branches. If the signal at the output of this transmitter were to be viewed using a traditional
oscilloscope, it would theoretically produce a perfect line at level one (assuming everything was
normalized to one). But in practice, only small dips would be seen along the way (as shown in
the top right‑hand graph in Figure 22) due to the response time of the electrical components.
Unfortunately, this information is not adequate. Not only is it essential to know that there is a
phase shift, it is also essential to know what phase the transition is in to be able to decode QPSK
data. This information is shown on the right‑hand side of the following graph.
π 0
Re (E)
MZM 3π/2
Time
CW Laser
3π/2
MZM π/2 π/2 Im (E)
π
π 0
Phase
Re (E)
π/2
Encoder/Driver
3π/2
0
Time
Figure 22. Block diagram of a DQPSK transmitter; intensity and phase diagrams of DQPSK transmitters.
Looking at the generic block diagram of a DP‑QPSK transmitter shown in the figure below, it is
easy to see several adjustment points that need to be properly calibrated:
• Four radio frequency (RF) drivers (RF1 to RF4)
• Two delays between the I and Q data branches (D1 and D2)
• Four Mach‑Zehnder modulator biases (B1 to B4)
• Two bias levels for the phase shifters used to create the quadrature signals (B5 and B6)
SERDES RF4
pre-coder RF RF3
10 x 11.2 G driver RF2
to RF1
4 x 28 G D1 D2
MZ1
CW laser B1
MZ2 π/2
B2 B5
MZ3
B3
MZ4 π/2
B4 B6
Figure 24. Block diagram of a DP‑QPSK transmitter.
Some of the most common impairments are based on the following incorrect adjustments:
• Quadrature errors (the phase error between the “I” and “Q” branches of the polarization)
• Poor modulator bias (will affect the opening of the eye diagram)
• “I” and “Q” gain imbalance (instead of having a square constellation, it should be
somewhat rectangular)
• Skew (the timing difference between the “I” and “Q” branches)
• Jitter‑related issues (data‑dependent or random clock jitter)
• Chirp
Clock
Recovery
Sx + LOx
Optical Balanced Decision/
Sx – LOx ADC/DSP
Input Detector FEC, Etc.
Signal Sx +jLOx
Balanced Decision/
90°- Sx –jLOx ADC/DSP
Detector FEC, Etc.
Polarization
Division
Sy + LOy
Optical Balanced Decision/
Hybrid Sy – LOy ADC/DSP
Detector FEC, Etc.
CW Local
Sy +jLOy
Oscillator
Balanced Decision/
Sy –jLOy ADC/DSP FEC, Etc.
Detector
It is very difficult to distinguish between impairments generated at the transmitter and impairments
generated by the actual receiver itself. As mentioned above, this has a lot to do with the software
and the receiver. The data provided by the coherent receiver only provides the effective number
of bits and bandwidth needed to recover the signal, but is not enough to provide a sufficient
evaluation of the behavior of the receiver itself.
DSP algorithms could be evaluated by simulating impairments on a link. For example, chromatic
dispersion (CD), polarization mode dispersion (PMD) and optical signal‑to‑noise ratio (OSNR)
could be generated using emulators and enabling evaluation of the DSP algorithms.
However, DSP algorithm qualification would not need to be performed in manufacturing, because
if qualification passes the first time, it will also pass in all other cases. For instance, it is important
to determine the capacity of compensation for CD and PMD, in addition to how it will behave,
because it will always behave in the same manner. As such, DSP algorithm qualification is most
likely only required at the research stage prior to final system validation.
The receiver can therefore be completely characterized by using a golden transmitter to test its
integrity, and then comparing the BER or Q factor before and after equalization.
Unfortunately, many other instruments would require a similar procedure. Like golden transmitters,
emulators (CD, PMD and OSNR) are stand‑alone equipment; they do not exist in a single,
calibrated entity.
40G Media Independent Interface (XLGMII) 100G Media Independent Interface (CGMII)
40G Attachment Unit Interface (XLAUI) 100G Attachment Unit Interface (CAUI)
40GBase-SR4/LR4/KR4/FR 100GBase-SR10/LR4/ER4/CR10
Lane Minimum Wavelength (nm) Nominal Wavelength (nm) Maximum Wavelength (nm)
L1 1520 1523 1526
L2 1528 1531 1534
L3 1536 1539 1542
L4 1544 1547 1550
L5 1552 1555 1558
L6 1560 1563 1566
L7 1568 1571 1574
L8 1576 1579 1582
L9 1584 1587 1590
L10 1592 1595 1598
IEEE P802.3bm next‑generation 40GE and 100GBASE‑SR4: IEEE established the P802.3bm
task force in 2012. One of their goals is to define four‑channel 100G PMA‑to‑PMA chip‑to‑chip
and chip‑to‑module four‑channel retiming electrical interfaces. This means CAUI‑4 based on
4X25G will pave the way for the production of CFP4. In addition, this project will define 40GE
40 km singlemode fiber 40GEBASE‑ER4 and 100GE four‑channel 100 m multimode fiber
100GBASE‑SR4.
Figure 28. The EXFO’s 88000 Power Blazer product family integrates a number of physical‑layer diagnosis tools.
The first picture shows the Rx power level of each optical lane.
The second screenshot shows the frequency offset for each physical lane.
The optical interface is based on parallel optics transmission, through which a number of
wavelengths (usually four or ten) is combined on the same fiber pair to form the 100G rate.
This functionality is handled by the CFP optics at both ends of the link. This is very different
from 10G transmission, which uses serial transmission with a single wavelength to carry the
full line rate. At the moment, the technology behind the pluggable SFP+ and XFP optics used
for 10G interfaces is more mature and stable, resulting in a product that is extremely reliable
and quite inexpensive. In contrast, 100G CFP technology, which is much more complex than
10G SFP+/XFP, is still evolving and does not yet produce consistently reliable optical interfaces.
Based on our experience with multiple carriers around the world, the likelihood of having a
faulty or unreliable CFP is quite considerable. If this is not detected by the carrier during link
turn‑up, it will result in significant schedule delays, or worse, poor link performance. Furthermore,
the cost of CFP interfaces today remains high and their availability is still somewhat limited.
Consequently, it is neither practical nor economically feasible for carriers to keep spares of these
CFP optics readily available for all 100G links and to replace them when they are suspected of
poor performance.
In order to ensure the proper deployment and optimal performance of 100G links, it is imperative
that carriers use the right tool to test the stability and reliability of every CFP being used.
The following section will present EXFO’s 100G test solution and its unique set of functionalities
specifically designed for CFP qualification, making it the tool of choice for 100G field deployment.
In addition to detailing the benefits of such tests, it will highlight other tests for lab qualification.
In cases where certain errors appear intermittently on the network, making the link unstable,
CFP Health Check is able to quickly identify these errors thanks to its easy‑to‑use graphical
user interface and suite of applications. Besides the above‑mentioned CFP laser, power and
frequency test, CFP Health Check includes the processes described below.
The MDIO section can be used to read the CFP4 temperature, enable advanced CFP4 functions
and even set the CFP4 to troubleshooting mode.
The next section examines some of the tests that can be performed automatically, using the
CFP Health Check application−eliminating the need for manual intervention while minimizing
the chance of error. The CFP Health Check application menu (shown in Figure 30) supports
predefined, but configurable OTN and Ethernet tests; one
or several tests can be selected at once, thereby optimizing
and reducing test time.
c. Skew Testing
100GBASE-R PCS
The CFP stress‑testing troubleshooting tool’s skew function
FEC1
(shown in Figure 31) can be used to inject different skew
PMA (20:10)
levels into one or multiple lanes. With the ramp function, these
CAUI
different tests (whether alternate or lane‑per‑lane) help verify SP1 SP6
PMA (10:n)
the far‑end receiver’s buffer tolerance level to skew, based on
the standardized skew points of the 802.3ba standard. SP2 SP5
PMD
SERVICE
PMD INTERFACE
Skew Ramp on Single PCS Lane SP3 SP4 MDI
1500 MEDIUM
Skew (bits)
1000
100GBASE-R
500
0
PCS or LL
Test
Iteration
CAUI Lane 0 1 2 3 4 5 6 7 8 9
PRBS Pattern 1 31 23 23 23 23 23 23 23 23 23
PRBS Pattern 2 23 31 23 23 23 23 23 23 23 23
PRBS Pattern 3 23 23 31 23 23 23 23 23 23 23
PRBS Pattern 4 23 23 23 31 23 23 23 23 23 23
PRBS Pattern 5 23 23 23 23 31 23 23 23 23 23
PRBS Pattern 6 23 23 23 23 23 31 23 23 23 23
PRBS Pattern 7 23 23 23 23 23 23 31 23 23 23
PRBS Pattern 8 23 23 23 23 23 23 23 31 23 23
PRBS Pattern 9 23 23 23 23 23 23 23 23 31 23
PRBS Pattern 10 23 23 23 23 23 23 23 23 23 31
Figure 32. PRBS‑pattern crosstalk test table.
EXFO’s Transport and Datacom 40G/100G solutions now include CFP Health Check
capabilities, which can be used to validate the entire network, making it extremely valuable to
100G link deployment. The modular FTB‑2 Pro/FTB‑4 Pro platforms also include optical physical
tools, specifically a visual fault locator and a fiber inspection probe, providing field technicians
with everything they need to commission 40G/100G networks in a single portable solution.
Validation tests of the CFP/CFP2 are still required before each deployment in order to analyze the
main status of the CFP/CFP2, and ensure that there are no errors. In many cases, the CFP/CFP2
is considered the weakest link in the network; in such cases, it is critical to ensure that the
CFP/CFP2 is completely functional prior to deploying 100G.
EXFO’s CFP‑to‑CFP2 adapter, which can be inserted into a CFP interface, allows field technicians
to validate CFP2 transceivers before deployment, thus minimizing risk.
EXFO’s FTBx‑88200NGE Power Blazer and FTB‑890/890NGE NetBlazer solutions now provide
support for CFP4 and QSFP+/QSFP28 pluggable interfaces. These interfaces offer higher port
density, require less power and are much more cost efficient. For example, a CFP4 transceiver is
half the size of a CFP2 transceiver, a major improvement. QSFP interfaces are more suitable for
short reach applications, with reaches below 10 km, making them attractive for Web‑scale data
centers, where cost per bit is an important factor. The FTBx‑88200NGE Power Blazer includes
CFP Health Check support for CFP4 and QSFPP+/QSFP28.
As a result, data centers and service providers around the globe are experiencing massive
growth in data communications through their infrastructures, which are constantly evolving to
support current and future traffic needs. The evolution of technology is naturally shaped by the
challenges inherent to the industry. Network devices are therefore designed in light of important
aspects, such as equipment size, power consumption and cost. One component that has a
direct influence on all of these three aspects is the transceivers used in network devices. These
transceivers, also known as pluggables, vary based on the data rates that they transmit and
receive. Optical transceivers have evolved from SFP to SFP+, XFP, CFP, CFP2, CFP4, QSFP+
and QSFP28. The following table provides a description of each one of these transceivers and
the rates at which they operate:
The technology used in each one of these pluggables is complex and constantly evolving with
the arrival of new manufacturers. Moreover, each type of transceiver is at a different maturity
level. These devices are often considered the weakest component of communication links and
therefore require special attention. Some, particularly those operating at rates below 16 Gbit/s,
have reached low price levels, and for this reason, some service providers prefer to just swap
devices whenever something appears to not be working as expected. This is a risky approach,
because some devices may only be borderline good, meaning that after deployment and service
activation, their performance may fall below the expected thresholds.
The cost of devices operating at 40 Gbit/s and 100 Gbit/s is higher. Web‑scale companies
and service providers cannot afford to have a significant number of spare units, and the impact
of faulty links at higher rates is certainly more important. It is also worth mentioning that these
devices (especially the CFP4 and QSFP28) are new to the market, and therefore their reliability
level has not yet achieved an acceptable standard. For all these reasons, it is extremely important
to acquire and use tools that quickly and accurately assess the quality of these pluggable
transceivers in order to ensure that deployment and service activations are done right—in a single
test execution and a reasonable amount of time. This will reduce the risk of future failures and
the need to perform costly troubleshooting activities later on.
The test uses a physical fiber loopback to validate the device signal continuity, integrity and
excessive skew generation. Appropriate signal attenuation should be used to simulate network
condition and protect the optical device.
Data Centers
Web-scale companies have a myriad of ports in their data‑center farms and are gradually
deploying hardware equipped with QSFP28 ports that operate at 100G. The quality and maturity
of these devices is linked to their development, with issues related to shutoff lanes, power levels
and control input/output (I/O) interfaces often arising.
The intelligent iOptics test application enables data centers to quickly validate pluggable optical
transceivers and meet the aggressive deployment schedules that are a part of today’s business
environment.
NEMs
NEMs are continuously introducing new hardware and software features to their designs.
As such, their engineers must ensure that all optics are reliable during the design verification
phase. NEMs cannot afford to jeopardize the validation of their design through use of a faulty or
non‑reliable transceiver, and therefore, proper operation of all pluggable transceivers is critical
and must be validated.
Before the sequence of tests is initiated, the iOptics test application performs a basic module
validation by collecting key device information such as vendor, module type, module ID, part
number and speed (see Figure 37). This enables the user to visually confirm that the desired
device is present in the selected port. The basic module validation also assesses what type of
test the optical device under test can support.
The series of tests is as follows, with the test application fulfilling two tasks: the monitoring and
execution of each subtest.
3) Temperature Monitoring
The Temperature Monitoring test determines the internal temperature of an optical device in
degrees Celsius until the end of the test. A pass verdict is obtained if the maximum temperature
is equal to or less than the temperature threshold over the entire duration of the test.
A fail verdict indicates that the pluggable is not operating at the range specified by the
manufacturer.
The user can configure the duration of the test, and a passing verdict will be assigned to the test
if the following conditions are met:
• No loss of signal (LOS)
• No pattern loss
• Bit error count ≤ BER threshold
For devices operating at multiple lanes, if any lanes present a problem, this will result in a fail
verdict.
The skew threshold is also automatically configured by the iOptics test application based on the
rate/protocol used. A pass verdict is assigned to the test when the highest measured skew is
equal to or less than the skew threshold.
In the event of a fail verdict during the execution of any subtest, the test is aborted and the fault
is reported to signal that the faulty transceiver device should not be used.
Once the configuration has been completed, the user can start the test, which will bring up the
Summary page containing the key results and verdicts associated with each phase in the test
sequence, as well as monitoring tasks related to each phase. A progress status is provided for
each test. It is worth noting that several tests are executed quite quickly.
The FAS uses the first six bytes in row one and columns one to six. G.709 uses FAS to provide
framing for the entire signal and to identify out‑of‑frame (OOF) and loss‑of‑frame (LOF) conditions.
The MFAS: G.709 supports the multiframing structure, in which some of the OTUk and ODUk
overhead signals could span multiple OTU frames. Examples are the trail trace identifier (TTI)
and tandem connection monitoring activation (TCM‑ACT) overhead signals. A single MFAS
byte is used to extend command and management functions over several frames. The MFAS
byte is defined in row one and column seven of the G.709 frame, and is incremented for each
OTUk/ODUk frame, providing a 256 multiframe structure.
2
Payload OTU
3 ODU (TCMi)
FEC
OH
4
Byte 1 2 3 4 5 6 7
1 FAS MFAS
OTU Overhead
The OTU overhead is composed of the section‑monitoring (SM), general‑communications‑channel
(GCC0) and reserved (RES) bytes, as shown in the figure on the next page.
• SM bytes: These are comprised of the TTI, bit‑interleaved parity‑8 (BIP‑8), backward error
indicator (BEI), backward incoming alignment error (BIAE), backward defect indicator (BDI),
and incoming alignment error (IAE).
• SM‑TTI: This is a one‑byte overhead field defined to support 64‑byte trace signals. TTI is used to
identify a signal from the source to the destination within the network. The TTI includes the so‑called
access‑point‑identifiers (API) fields, which are used to specify the source access point identifier
(SAPI) and destination access point identifier (DAPI). The APIs include information regarding the
country of origin, network operator and administrative details.
• SM error BIP‑8: This is a one‑byte, error‑detection code signal. The OTUk BIP‑8 is computed over
the OPUk area of a specific frame and inserted in the OTUk BIP‑8 overhead two frames later.
• SM‑BDI: This is a single‑bit signal defined to convey the signal fail status detected in the upstream
direction.
• SM‑BEI and SM‑BIAE: This is a four‑bit signal used to convey (in the upstream direction) the
number of interleaved‑bit blocks detected in error by the section monitoring BIP‑8 code. This signal
is also used to convey (in the upstream direction) an IAE condition that is detected in the section
monitoring IAE overhead.
• The GCC0 field, which resembles the data communications channel (DCC) in SONET/SDH, is
currently undefined. However, it will likely be used for functions such as network management
or control plane signaling for a protocol (e.g., generic multiprotocol label switching [G‑MPLS]).
• The RES fields, found throughout the overhead, are set aside for future use.
The ODU overhead supports two classes of ODUk maintenance signals, reported using
path‑monitoring‑overhead (PMOH) status (STAT) bits and tandem connection monitoring (TCM)
STAT bits. Through either PMOH or TCM STAT bits, the following ODU conditions can be reported:
alarm indication signal (ODUk‑AIS), open connection indication (ODU‑OCI), locked (ODUk‑LCK),
and generic AIS. In addition, the ODUk overhead supports automatic‑protection‑switching (APS)
functionality. The ODUk overhead is broken into the following fields: RES, PM, TCMi, TCM ACT,
FTFL, EXP, GCC1/GCC2 and APS/PCC.
2
Payload OTU
3 ODU (TCMi)
FEC
OH
4
Byte 1 2 3 4 5 6 7 8 9 10 11 12 13 14
1
2 PM & TCM
RES TCM6 TCM5 TCM4 FTFL
TCM ACK
3 TCM3 TCM2 TCM1 PM EXP
TTI BIP-8
1 4 5 6 8
0 Applies to PM
SAPI BEI/BIAE BDI RES and TCM1-6
15
16
DAPI
31
32 Operator-
63
Specific
• RES bytes are undefined and are set aside for future applications.
• Path monitoring (PM) enables monitoring of particular sections within the network, as well as
fault locations in the network. The PM bytes are configured in row three, columns 10 to 12, and
include subfields similar to those in SM, including TTI, BIP‑8, BEI, BDI and STAT subfields.
• The tandem connection monitoring (TCMi) fields define six ODU TCM sublayers. Each
TCM sublayer contains a TTI, BIP‑8, BEI/BIAE, BDI and STAT subfield associated with a
TCMi level.
• Tandem connection monitoring activation/deactivation (TCM‑ACT) is a one‑byte field
located in row two and column four. TCM‑ACT is currently undefined in the standard.
2
Payload OTU
3 ODU (TCMi)
FEC
OH
4
Byte 15 16 17 24
1 RES JC
2 RES JC
3 RES JC
4 PSI NJO PJO
0 PT
1
2
PSI
CSF
ı RES
ı
255
The following figure illustrates the process in which the FEC protocol interleaves one overhead
byte and 238 data bytes to compute 16 parity bytes, thus forming 255‑byte blocks, i.e.,
the RS (255,239) algorithm. The key advantages of interleaving the information are reducing
the encoding rate of each stream relative to the line transmission rate, and reducing the sensitivity
to bursts of error. The interleaving, combined with the inherent correction strength of the
RS (255,239) algorithm enables the correction of transmission bursts of up to 128 consecutive
erroneous bytes.
OTU
Sub-Rows
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
1
2
OTU
Rows 3
4
1..............................239. 240...........255
OPU OH
.. 2 OPU Payload OTU FEC
.. (4 x 256 bytes)
.. 3 ODU OH (Client Signal)
.. 4
..
.. 1.................................................................................14. 15. 16. 17..............................32........................................3824. 3825........................4080
Row 1
Sub-Row 1 OTU
(codeword) Sub-Rows
Sub-Row 2 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
(codeword)
1
Sub-Row 3 2
(codeword)
OTU
Rows 3
OH Payload Bytes FEC 4
1..............................239. 240...........255
To determine whether the FEC algorithm is functioning correctly, generate some correctable FEC
errors with the test equipment and verify whether the errors can de detected and then correct
any bit errors. The signal received by the test instrument should be free of bit errors. You can
also use the instrument to generate uncorrectable errors, following which the system under test
should be able to properly report the alarm to the network administrator.
Finally, use the instrument to run the FEC pressure test by generating a random bit error in OTU
frame. The device under test (DUT) should be able to correctly identify all the bit errors and
return an error‑free OTN signal.
G.70
9
NE
Figure 46. Connection diagram of response test using EXFO’s FTB‑2 Pro.
The figure below shows common stimulus signals, along with corresponding upstream and
downstream responses of the DUT:
ODU multiplex functionality testing is also a key parameter that needs to be validated as part of
G.709. In order to determine the appropriate multiplexing capability of the network element under
test, the test equipment is used in OTN‑decoupled mode to generate a lower‑rate signal on the
transmit side. The transmitted signal then gets multiplexed within a higher‑order ODU signal on
the G.709 network element, along with the proper overhead and FEC bandwidth to compose
the final OTU signal. Finally, the received OTU signal is checked at the test equipment in order
to verify the proper multiplexing with the proper frequency justification and synchronization,
as shown in Figure 48 below.
ODU multiplexing structures are identified by the OPU multiplex‑structure‑identifier (MSI) bytes.
Figure 49. Client signal mapping test configuration: both the transmitting and receiving ports require the decoupling mode.
2.8. Packet/Ethernet over OTN Test: ODU0, GMP ODUflex and Hitless
Adjustment of ODUflex (GFP) (HAO)
The Ethernet, which was responsible for the first ever successful deployment of OTN devices
for SDH/SONET signal transport, continues to break the application limit of the enterprise
network realm and extend towards the public network realm. In response, the International
Telecommunication Union (ITU) has done a lot to support Ethernet by defining and standardizing
the transport of multirate Ethernet signals over the OTN network, including Gigabit Ethernet,
10G Ethernet, 40G Ethernet and 100G Ethernet.
ODU0
One important concept for Ethernet support is to define a
transmission container with a suitable size for Gigabit Ethernet.
In the initial definition of OTN, the optical channel (OCh) data unit
ODU1 is the smallest transmission container, exclusively used
to transport single STM‑16 signals with a payload capacity of
2 488 320 kbit/s. This also signifies that if the ODU1 is merely
transporting single Gigabit Ethernet signals, a lot of bandwidth will
be wasted. Hence, the ITU defines the ODU0 as half the payload
bit rate of the OCh payload unit OPU1, i.e., 1 238 954.310 kbit/s.
Figure 50. GFP‑T overhead‑byte analysis interface on EXFO’s FTBx‑88200NGE OTN analyzer.
ODUflex is able to provide flexible containers of variable sizes and client signals of variable
sizes, using a principle similar to virtual concatenation (VCAT) in multiservice‑transport‑platform
(MSTP) technology. ODUflex offers an efficient yet simple way to map variable‑rate packet
services or fiber‑channel‑like constant rate services to ODU. ODUflex employs a 1.25G tributary
slot (ODTUGk) to create a variable container, thus ensuring that the client signal is mapped
into this container and then transported via the ODUk signal. For non‑CBR service signals,
ODUflex employs the frame‑mapped generic framing procedure (GFP‑F) to map signals. The main
advantage of ODUflex is that it enables unoccupied slots to be reused and provides similar
bandwidth‑variable block‑oriented‑device (BOD) services.
Hitless adjustment of ODUflex (HAO) is a scheme for hitless adjustment of the bandwidth of an
ODUflex (GFP) connection in an OTN network, and is based on the G.709 recommendation.
Through the HAO protocol, the system is able to increase or decrease the client signal data
rate carried by the ODUflex (GFP) over the end‑to‑end path. The process closely resembles the
dynamic link capacity adjustment scheme (LCAS) for adjustment of VCAT in the MSTP system.
However, it should be noted that the HAO requires the involvement of all the nodes on the entire
path, as opposed to just two ends (which is the case for VCAT, a merely logical concatenation).
In general, every VCAT tributary can have different paths, with variable time delay differences
being offset on the receiving end with cache. However, HAO requires all ODUflex tributaries to
go along one path, thus addressing the significant need for cache that affected LCAS in the past.
HAO overhead is mainly implemented by resize control overhead (RCOH). RCOH is located in
the HO OPUk tributary slot overhead (TSOH) and the OPUflex overhead. RCOH mainly consists
of two parts, i.e., link connection resize (LCR) control protocol and bandwidth resize (BWR)
control protocol, respectively.
Figure 53. ODUflex overhead byte analysis interface on EXFO’s OTN analyzer.
In an OTN network, the packet Ethernet signal will be the main service type. The GE/10GE/
40GE/100GE Ethernet signal can be mapped to OPU0/1/2/3/4/flex and OPU1/2/3‑X by means
of GFP‑F. In addition, the 40GE Ethernet signal can be mapped to OPU0/3 by the generic
mapping procedure (GMP). The 10GE signal can also be mapped to OPU2e by the BMP.
The 100GE signal is restored as 64B/66B codes by the GMP and is mapped to OPU4.
A more flexible or more generic method exists, namely, the GMP. In all cases (e.g., the maximum
frequency offset of the client signal in ppm and the minimum frequency offset of the server signal
in ppm), this method is able to map any client signal rate into a corresponding payload container
as soon as the server signal rate is determined to be higher than the client signal rate.
In the process of mapping the 40G/100G Ethernet client signal into OTU3/OTU4, where the service
frame (or multiframe) can contain data or stuffing information, the Sigma‑Delta data/stuffing algorithm
is used to distribute the data as evenly as possible within the service frame. During this process, to
guarantee that the client signal is restored completely without any demapping errors, the receiving
end must be aware of the data stuffing method for every frame/multiframe at the transmitting end.
In the GMP, the OPUk overhead contains the PSI, PT, and six justification bytes (JC1, JC2,
JC3, J4, J5, and J6) as the justification‑control bytes of the GMP, as shown in the figure below.
Bytes JC1, JC2 and JC3 contain 14‑bit Cm identifiers (C1, C2...C14), an increment indicator
(II), decrease indicator (DI), and an 8‑bit error‑check CRC‑8 that is used to check bytes JC1,
JC2 and JC3 for errors.
The GMP uses these justification‑control bytes to notify the receiving end of the number of bytes
in the OPUk payload, and this process will be carried out in every frame. The size of the payload
bytes is represented by the Cm value, whereas the variation in the byte size of every OPU frame
is represented by two specific indicators, the II and the DI. Bytes JC4, JC5 and JC6 contain the 10 bit
CnD byte (D1, D2 …D10), bytes four through eight are 5‑bit error‑check CRC‑5, and byte nine is
reserved (RES) for later use. The CnD parameter is used in the GMP, providing a scheme that can be
used to carry some client signals with more stringent restrictions on jitter, e.g., SONET/SDH signals.
Overclocked OTN compensates for the rate mismatch between 10 GigE LAN and the OPU2
payload by raising the overall OTU2 data rate from the standard 10.709 Gbit/s to fit the 10 GigE
LAN client signal. Obviously, this modification of the standard OTN line rate will bring about
interoperability issues, and the option for aggregating OTU2 signals into OTU3 will be lost;
however, ODU3e does allow for multiplexing. On the positive side, overclocked OTN offers real
bit transparency of 10 GigE LAN signals—a necessity for the mass deployment of 10G services.
Overclocked OTN supports OTU1e, OTU2e, OTU3e1 and OTU3e2 optical line rates for mapping
of 10 GigE LAN signals. Furthermore, OTU1f and OTU2f line rates are used for mapping Fibre
Channel signals. The following figure illustrates an overclocked OTN application scenario.
The transparent transportation of 10 GigE LAN signals means that the full 10 GigE data rate
(10.3125 Gbit/s) is transported over OTN, including the PCS 64B/66B coded information,
inter‑packet gap (IPG), MAC FCS, preamble, start of frame delimiter (SFD), etc. The OTN
clocking is derived from the Ethernet customer signal (±100 ppm) rather than that of a standard
OTU2 signal (±20 ppm). Therefore, overclocked OTN is mainly used for point‑to‑point data paths.
The OTN overclocked interface rates and corresponding client rates are listed in the table below:
Figure 58. OTN overclocked interface rates and corresponding client rates.
The signal rate is different from that of the previously defined OTU2, and the PT value is “0x80”
in the OTN frame. OTU1e is similar to OTU2e. The only difference is that it does not contain any
fixed stuff bytes, and therefore, the rate is slightly lower. The diagrams are as follows:
3824
3825
4080
4
3824
3825
4080
080
14
15
16
17
14
15
16
17
1 78 1 78
1 1
FAS OTU FAS OTU
OPU OH
OPU OH
2 OH Fixed 2 OH
OPU Payload Stuff OPU Payload FEC OPU Payload FEC
3 ODU OH 3 ODU OH
Bytes
4 4
Figure 59. OTU2e with fixed stuff bytes. Figure 60. OTU1e without fixed stuff bytes.
Ethernet
Mapper 40G/100G
OTN Ethernet
Ethernet Device
Demapper
Using EXFO’s FTBx‑88200NGE Power Blazer, service providers can map 40G/100G Ethernet
services over OTN with different traffic characteristics, run bit‑error‑rate (BER) tests across
the OTN, and measure the ratio of the error bits compared to the number of sent bits. In this
testing configuration, the FTBx‑88200NGE Power Blazer module provides complete analysis
of the OTU3/OTU4 (including the OPU, ODU, OTU and OTL layers), related alarms, errors
and skew measurement. The 40G/100G module also provides GMP‑related measurements,
including Cm and CnD statistics, to ensure proper recovery of the client signal at the receive
end. Once the OTN with the embedded 40G/100G Ethernet client service is qualified from end
to end, service providers can take their service validation a step further by qualifying the actual
40G/100G Ethernet service from client to client. This includes validating the 40G/100G Ethernet
IP traffic transmission with 100‑percent throughput, and ensuring that latency measurement is
not impacting customers’ SLAs.
Layer Defect
OTL LOF, OOF, LOL, LOR, OOR, Inv. Marker, FAS
FEC FEC‑CORR, FEC‑UNCORR
OTU AIS, LOF, OOF, LOM, OOM, BDI, IAE, BIAE, BIP‑8, BEI, FAS, MFAS
ODU AIS, OCI, LCK, BDI, BIP‑8, BEI, FSF, BSF, FSD, BSD
OTU AIS, CSF
Pattern Pattern Loss, BER
Figure 64. Common defect triggers.
PASS FAIL
Service Disruption
Threshold
Time
Service Disruption
Verdict Threshold 50.0
Longest Shortest Last Average Total Verdict
Count
(ms) (ms) (ms) (ms) Duration (ms)
4 PASS
15.462 0.419 1.219 4.432 17.730
To meet mesh testing requirements, EXFO’s OTN tester is able to run the service disruption test
for all lanes, with granularity as small as ODU0.
The test interface and results are shown in the following figure:
Figure 71. Testing of TTI SAPI/DAPI/operator information using the FTBx‑88200NGE Power Blazer.
TCM1
Figure 72. TCM allocation.
Figure 74. Threshold values of the errored block number in one second for determination of the SES.
In‑Service Monitoring
In‑service monitoring is used to monitor each block by way of the internal BIP‑8 error detection
code (EDC). Each EDC bit is physically isolated from the block it monitors. It is usually impossible
to determine whether the error is in the block or its controlling EDC bit. If there is any discrepancy
between the EDC and its controlled block, the controlled block is always deemed erroneous.
With a device tester or OTN tester, the operator can monitor the channel performance and status.
Out‑of‑Service Measurement
The block‑based bit error test can also be out of service, which is also the basis for a bit error
test with an OTN tester. As a basic requirement of the engineering test, the capability of the
out‑of‑service bit error test must be stronger than it is online.
With EXFO’s tester, the user can also set ten user‑defined patterns. In addition to the PRBS,
different kinds of user information signals (such as GE, 10GE, FC or SDH signals) can be
mapped into OUT for testing. The bit error test normally requires long‑term, error‑free verification,
e.g., a 24‑hour or 72‑hour error‑free test.
Let’s have a look at points to keep in mind while testing OTN and also the key points addressed
by enhanced multichannel OTN with mixed mapping.
Connectivity validation
Different methodologies are available to validate connectivity. From a physical perspective, the
goal is to generate a test signal to the device under test (DUT) and analyze the response:
Scenario 2: Both the FTBx‑88200NGE and the DUT are connected to the network
Scenario 3: Multiple DUTs are connected to the network with a single test apparatus that
generates the stimulus. In the following example, the FTBx‑88200NGE is connected to the
network via an OTU4 link. The interoperability monitoring of multiple DUTs is done via OTU2 links.
The following table shows the maximum number of channels for carrying the OTU4 signal:
Multiplexing Number of channels
OTU4/ODU4/ODU0 80
OTU4/ODU4/ODU1 40
OTU4/ODU4/ODU2 10
OTU4/ODU4/ODU3 2
OTN’s multiplexing capabilities allow it to carry more data containers inside an ODU4. Users
wishing to test network element multiplexing functionality more exhaustively can use multichannel
OTN enhanced with mixed mapping to test a basic scenario by saturating the pipe with the same
type of data units. This is equivalent to testing the capacity of a highway with the same‑sized cars
(testing scenario 1 in Figure 76). An example of such a test would be to configure 80 ODU0
channels on an ODU4 to be transported on an OTU4, as in Figure 77.
Figure 76. Multichannel OTN enhanced with basic mixed mapping scenario
Once the channel configuration is complete, the test page displays a summary of the ODU
container types configured as well as the quantity of channels used and any unused capacity,
helping the user to graphically confirm the structure to be tested (Figure 81).
To calculate disruption time, the application considers a specific defect, a no‑defect time of 300
ms, and an SDT threshold value provided by the user (Figure 83). SDT duration is calculated
as the time between the detection of the first defect and the end of the last defect, excluding
the no‑defect time.
Let us suppose that the defect type is set to AIS and a no‑defect time configured to 300 ms.
The first AIS event lasts 20 ms, followed by an absence of the alarm for 150 ms and another
AIS event lasting 45 ms. No other alarm is observed. Since the no‑defect time is set to 300 ms,
the 150 ms interval is insufficient for the system to clear the service disruption. Therefore, the
total SDT is 215 ms (20 + 150 + 45), with one single disruption event counted (Figure 84).
If the user were to select the pass/fail verdict for the SDT measurement and define the SDT
threshold as 50 ms, the first scenario would result in a failing verdict with an SDT measurement
of 215 ms.
Per‑channel testing and validation with this test application makes it possible to easily measure
SDT on every channel in order to detect defective channels during turn‑up or advanced
troubleshooting. Overall test time is reduced because each channel is tested individually,
providing its own statistics.
The interface offers a simple and unique monitoring system that detects channel alarms and
errors in real time, indicating with a specific color code the status of every channel in the test
configuration (Figure 86). Green indicates no channel errors, orange shows the channel is
selected, yellow indicates some alarms or errors have been recorded since the test started, and
red signals that there is currently an active error or alarm on that specific channel.
Figure 86. Status of every channel in the test configuration is color coded
Lastly, to make absolutely sure that the DUT is detecting and reacting correctly to the alarms
and errors (e.g., BIP‑8, AIS) that have been injected, users can inject those alarms and errors
in a very intuitive way—using only a single button, either on a specific channel or on all channels
included in the configuration.
BDI
TTI TTI undefined.
BIP-8 REScertain functions,
IAE
31 SONET/SDH,
DAPI and it is currently However, it will likely be used for
32 such as network management or control plane signaling for G‑MPLS
Operator- Byte protocols.
10
63 Specific
1 2 3 4 5 6 7 8 9 10 11 12 13 14
MFAS
FAS SM GCCO RES
OPU OH
OTU 2 9 10 1 2 3 4 5 6 7
Frame 3 16 OH
ODU (Client Signal) (4 x 256 bytes)
BEI/BIAE
BDI
TTI TTI BIP-8
IAE
DAPI
4 31
32 Operator- Byte 10
General Communication Specific 1 and 2 are two‑byte
63 Channel 15 16 long fields 0inside PT the ODUk overhead
TCM 1
2 RES TCM6 TCM5 TCM4 FTFL 1
and are similar to the GCC0 field; GCC1 is located in row four, columns14 one and two; while
Mapping
ACT
3 TCM3 TCM2 TCM1 PM EXP 1 2 3 4 5 2 6 7 8 9 10 11 12 13
Mapping
GCC2 GCC2is located
APS/PCC in row four, RES columns three FASand3 four, at SM the ODU overhead. They are used to
MFAS
4 GCC1 GCCO RES
1 carry
2 3 end‑to‑end
4 5 6 7 client 8 9 information.
10 11 12 13 14 4 PSI
255
PM and TCMi (i= 1 to 6) 1 14-15-16-17 3824
1 2 3 1 FAS OH OTU OH
TTI BIP-8 OPU Payload
OPU OH
OTU 2 0 1 9-10 127
Frame 3 ODU OH Operator Operator- (Client Signal)
0 Forward FIF
PM 1 2 3 4 5 6 7 8 Identifier Specific
SAPI 4
15 BEI STAT
BDI
16 FTFL
DAPI Byte 3 128 129 137-138 255
31 TCMi 1 Backward FIF Operator Operator- 15 16 0 PT
32 2 3 4 5 6 7 8 Identifier Specific 1
Operator- TCM
BEI/BIAE 2 RES TCM6 TCM5 TCM4 FTFL 1
Mapping
STAT ACT
BDI
Mapping
4 GCC1 GCC2 APS/PCC RES 3
1 2 3 4 5 6 7 8 9 10 11 12 13 14 4 PSI
255
PM and TCMi (i= 1 to 6)
1 2 3
Recently, GCC channelsTTIhave been used
for different purposes, including synchronization and
BIP-8
encryption, and therefore, their integrity is prime for testing. 0 1 9-10
Forward FIF Operator
0 PM Identifier
1 2 3 4 5 6 7 8
The GCC BERT test 15
provides
SAPI a tool for measuring the bit error rate on these fields, providing statistics
BEI STAT
BDI
Customer SLAs dictate certain performance criteria that must be met, with the majority
documenting network availability and mean‑time‑to‑repair values that are easily verified. However,
Ethernet performance criteria are more difficult to prove, and demonstrating performance
availability, transmission delay, link burstability and service integrity cannot be accomplished
accurately by a mere PING command alone.
Portable RFC 2544 test equipment enables field technicians, installers and contractors to
immediately capture test results and demonstrate that the Ethernet service meets the customer
SLA. These tests also serve as a performance baseline for future reference.
Throughput Test
The throughput test defines the maximum number of frames per second that can be transmitted
without any errors. This test is done to measure the rate‑limiting capability of an Ethernet switch
as found in Carrier Ethernet services. The methodology involves starting at a maximum frame rate,
and then comparing the number of transmitted and received frames. If frame loss occurs, the
transmission rate is divided by two and the test is relaunched. If there is no frame loss during this
trial, the transmission rate is increased by half of the difference from the previous trial. Known as
the half/doubling method, this trial‑and‑error methodology is repeated until the rate at which
there is no frame loss is found.
The throughput test must be performed for each frame size. Although the test time during
which frames are transmitted can be short, the duration of the final validation must be at least
60 seconds. Each throughput test result must then be recorded in a report, using frames per
second (f/s or FPS) or bits per second (bit/s or BPS) as the measurement unit.
The test instrument sends traffic at maximum line rate and then measures whether or not the
network dropped any frames. If so, the values are recorded, and the test will restart at a slower
rate (the rate steps can be as coarse as 10%, although a finer percentage is recommended).
This test is repeated until there is no frame loss for three consecutive iterations, at which time
a results graph is created for reporting. The results are presented as a percentage of the
frames that were dropped, i.e., the percentage indicates the variable between the offered load
(transmitted frames) vs. the actual load (received frames). Again, this test must be performed
for all frame sizes.
Latency Test
The latency test measures the time required for a frame to travel from the originating device
through the network to the destination device (also known as end‑to‑end testing). This test can
also be configured to measure the round‑trip time, i.e., the time required for a frame to travel
from the originating device to the destination device, and then back to the originating device.
When the latency time varies from frame to frame, it causes issues with real‑time services.
For example, latency variation in VoIP applications would degrade voice quality and create pops
or clicks on the line. Long latency can also degrade Ethernet service quality. In client‑server
applications, the server might time out, or poor application performance could occur. For VoIP,
this would translate into long delays in the conversation, producing a “satellite call feeling”.
The test procedure begins by measuring and benchmarking the throughput for each frame size in
order to ensure that the frames are transmitted without being discarded (i.e., the throughout test).
This fills all device buffers, therefore measuring latency in the worst conditions. The second step is
for the test instrument to send traffic for 120 seconds. At the midpoint of the transmission, a frame
must be tagged with a time‑stamp and when it is received back at the test instrument, the latency
is measured. The transmission should continue for the rest of the time period. This measurement
must be taken 20 times for each frame size, and the results should be reported as an average.
Complementary Testing
Although not included in the defined RFC 2544 standard, another crucial measurement in
Ethernet networking is packet jitter. This measurement is critical, because excessive jitter
can cause major failures in real‑time applications such as VoIP or streaming video. In VoIP
applications, excessive jitter will cause dropout effects. In video applications, the image will
become choppy and pixelization could occur.
Service providers will usually benchmark their network and network elements in the lab and then,
from this knowledge, customize their RFC 2544 test routine for field turn‑up and installation.
For some, the number of frame sizes to test is too large, so they streamline their turn‑up testing
to include only the extremes, i.e., 64 and 1518 byte frames (could be longer if VLANs are used).
Some may keep all frame sizes, but decide to conduct only two tests, such as throughput and
latency. Others will decide to limit the number of tests and frame sizes. The critical consideration
is for service providers to comprehensively test before turning the circuit over to the customer,
because this will be the last time that they can fully access it without affecting subscriber services.
Because central‑switching POTN devices have the statistical characteristic of packet switching,
packet services need to be subjected to quality of service (QoS) classification. In other words,
higher‑demand services are classified as guaranteed service traffic (GST) in order to guarantee
that such services have no packet loss. Packet delay and packet delay variation (PDV) must be
provided with circuit‑switching characteristics for use in transporting higher‑priority services
such as the synchronous signal, real‑time service, VIP client special line, mobile backhaul, etc.
Low‑priority services are classified as the statistical multiplexed (SM) type, for which a best‑effort
strategy is adopted.
The diagram in the following figure represents a structural schematic of an SM‑based device.
The GSTs on the left are guaranteed (5 GE) signals. The one at the bottom right is an SM signal
(one 10G signal), and the one at the top right is of 10G output with an SM signal added.
1 Gbit/s
1 G1, 2 G1, 1 Max MTU
Figure 87. SM‑based schematic diagram where the GSTs on the left are guaranteed signals.
Test method: the EXFO FTBx‑88200NGE is used to simulate multichannel 100G signals.
This example shows four 100GE signals connected to the POTN NE1 that will aggregate them
to create a 400G signal. This 400G signal will be de‑multiplexed by the POTN NE2, allowing
for end‑to‑end testing. This validation is accomplished with 4 X FTBx‑88200NGE modules that
can support either CFP4 or QSFP28 100G interfaces. A similar scenario at 40G rate could be
accomplished. The user would configure four 10GE signals using the FTBx‑88200NGE with
the SFP+ ports, and test a 40G muxing application.
Figure 88. POTN aggregation and QoS test with EXFO’s LTB‑8 and FTBx‑88200NGE
Noisy Environments
Use of multiple lab equipment is only generating increased amounts of heat. For example, high‑
speed optical transceivers, such as CFP, CFP2, CFP4, need a very stable temperature to operate
and more powerful fans are required to reduce lab temperatures. This also applies to many
other network elements. Unfortunately, all the cooling fans being used tend to generate noise
that may disturb users located in the same area, making this environment a hard one to work in.
Space
Having adequate space in lab environments is a great challenge due to the different types of
technologies and pieces of equipment that must be extensively tested in the same facilities. As
a result, there’s not much leeway for users to move around or to be comfortable.
Remote Locations
Time to market is more crucial than ever and network equipment manufacturers are required to
have different teams working on the same project from various remote locations. Fast developing
strategies should not jeopardize security and teams from different regions of the world should
be able to have secure and efficient access to testing environments. There should not be any
risk incurred from sharing proprietary and confidential information when units are accessed
from remote locations.
Lack of Flexibility
Hardware equipment must be extensively tested in lab environments during long periods of
time; additionally, many users and testers must have access to different platforms and modules
at the same time; securing the utilization of specific cards or test sets becomes a complex, yet
fundamental, administration task.
GROUP A
Plat
f
Plat orm 1, Platform 1: Host Server
form Mod
5, M ule 1 (Hosting EXFO Multilink)
odu (GU
le 6 I)
(GU GROUP B
I)
Rem
o
Use te Up to 8 modules
rs per platform
Plat
f
Plat orm 2, AN
form Mod
2, M ule 3 er L
odu (SC
ustom Platform 2: Client
le 3 PI)
(GU C GROUP C
I)
Platform 3: Client
Platform 4: Client
Platform 5: Client New
FTB-2/4 Pro
Pla
tfo
rm
4,
Mo
du
le
1(
GU
I)
Module
Management
LINE SIDE
The following section details different deployment strategies. At this stage, the cards and systems
will be ready and fully tested during manufacturing. The next step is deployment, with the main
objective being to ensure that everything functions properly.
The system itself contains three main components: the transmitter, the link and the receiver.
The system’s performance is highly dependent on the DSP. Every vendor will have its own
unique proprietary test methodology, so it is important to define the actual parameters that will be
needed, and whether or not testing should be performed before or after deployment. The same
applies with regard to identifying and troubleshooting potential issues.
The range of offerings is therefore quite wide, depending on the quality of the DSP, the optics, the
local oscillator and other variables. While the OIF recommends a certain amount of compensation
per application, many carriers will not willingly deploy the most expensive DSP if they can make
do with a cheaper one.
So while in theory, an unlimited amount of CD can be compensated for, often a simple test for
CD will determine the amount that is residual on the link, which will in turn optimize the purchase
of compensation electronics.
Test Conditions
a. Greenfield
In greenfield deployments, which involve brand‑new fiber, typically only 100 Gbit/s coherent
signals will be carried. As such, the CD will be known, albeit not with high accuracy. However,
knowledge of the fiber type and length will provide an accurate assessment, making it easy to get
a good idea of the total CD per span and to select the proper DSP technology. Accordingly, CD
testing is not required in most greenfield applications.
XPM is a non‑linear effect (NLE) in which one carrier (wavelength) negatively influences a
neighboring wavelength. In OOK, only the amplitude of the transmitter is modulated.
0 1 0 0 1 1 1 0 1 0 0 1
Due to the power density, as this propagates (locally when a “1” is propagated), the medium
(glass) becomes locally heated. And although this local heating is very minute, it changes the
index of refraction (IOR) of the fiber very locally and propagates along the fiber, making it a
medium modulated in IOR.
Because OOK only uses amplitude information to detect the signal, this medium modulation
has no real impact. But, the IOR has a direct impact on the speed of propagation and, therefore,
on the phase of the modulated signal (currently most high‑end CD analyzers use phase‑shift
detection to measure the difference in group delay).
Accordingly, the OOK causes phase modification effects on a local scale, and the adjacent
transmission is phase‑encoded. As such, this phase modulation will have an impact on the phase
encoding, inducing phase noise known as XPM.
This would not be an issue in the presence of strong CD, because every wavelength travels at
different speeds and the medium modulation created by the OOK would not be time‑synchronized
(in‑phase) with the DP‑QPSK, so it would not create any XPM. But, as mentioned above, most
brownfield applications are optimized to result in very low residual CD, making these systems
prone to XPM.
In addition, the longer the fiber, the longer the duration of this interaction, resulting in more
significant XPM.
A practical solution for bypassing this problem is to use guard bands, e.g., freeing a few channels
between OOK and DP‑QPSK.
Figure 92. Adding a channel of a different modulation scheme may lead to NLEs.
… would require removal of a few neighboring channels, as shown in the figure below:
Although this does work, it is not extremely efficient. For example, adding a 100 Gbit/s channel
would require removal of two 10 Gbit/s channels, which would result in a bandwidth gain of
only 80 Gbit/s.
The noise created by XPM is not amplified spontaneous emission (ASE) noise and is therefore
not detectable by standard optical spectrum analyzers (OSAs). The OSA tests might indicate
sufficient OSNR, even if high BER is present. The only way to detect the presence of XPM is to
use EXFO’s WDM Investigator, which would indicate the presence of any XPM on the Non‑linear
Depolarization line, as shown in the figure below:
OSNR (dB)
14 14
13 13
12 12
11 11
10 10
-5 -4 -3 -2 -1 0 1 2 -5 -4 -3 -2 -1 0 1 2
Launch power (dBm) Launch power (dBm)
As power increases and channel spacing decreases, OSNR requirement increases. This is a
direct cause of XPM.
As discussed earlier, XPM occurs due to local heating of the transmission medium. It is therefore
normal that higher power creates more XPM, as shown in the graphs above.
However, power is not the actual limiting factor. The real issue is power density; the same
power on a smaller core will create more XPM than on a larger core. In commercial fibers,
G.652 dispersion‑shifted fiber has a larger core and is less prone to high power density, but is
also more frequently compensated for. Therefore, G.652‑based networks typically have lower
residual CD. Non‑zero dispersion‑shifted fiber (specified in ITU‑T G.655) has a much smaller
core, but because it naturally has lower CD, it is frequently not compensated for. For this reason,
the residual CD can be slightly higher, which is offset by the higher power density. Some G.655
fiber has a larger effective area than other G.655 fiber and, as a result, could be less prone to
this issue.
Accordingly, fiber type must be taken into consideration prior to creating a channel allocation plan.
In greenfield, this is not an issue. In the brownfield environment, the fiber type is often unknown.
Tools such as EXFO’s FTB‑5700 indicate the CD, in addition to measuring the length and
coefficent. Accordingly, they are ideal for the fiber‑type identification step required prior to
wavelength allocation on brownfield.
Once the allocation is complete, the WDM Investigator (which is pictured above and available
as an optional feature on EXFO’s FTB‑5240S‑P OSA) will show any residual XPM.
c. Mesh Network
Metro mesh networks, which include ROADMs, add flexibility, optimize bandwidth, allow for faster
disaster recovery, and allow wavelengths from different transmitters and different systems to be
routed differently as time goes by.
When these networks were deployed (with OOK in mind), it was important for each drop port
of the ROADM to be dropping a well‑compensated wavelength. At the network design stage,
each ROADM node was equipped with internal DCMs, each compensating sufficiently for the
length and type of fiber between itself and the previous ROADM.
Figure 97. Mesh network, short individual spans, and long total transmission distance.
As such, qualification of residual dispersion on a per‑span basis and fiber‑type information are
two critical pieces of information that the FTB‑5700, for example, can retrieve from this type
of network.
d. Raman Networks
Considering that the main predictable limitation of DP‑QPSK and coherent detection is
OSNR, the extra distance and amplification needs to be obtained with low‑noise amplification
piggybacked with the current amplifiers. This also needs to distribute the gain to lower the risk
of NLEs. Raman amplification is therefore very popular in most coherent systems, whether they
are greenfield or brownfield.
Fiber Span Raman
While erbium‑doped fiber amplifiers (EDFAs) use
a specially‑doped fiber as an amplification medium, Signal Pump
Now of course, like all amplifiers, the power of the pump can be adjusted. In counter‑propagation
schemes like the one highlighted in Figure 87, gain (in dB) is proportional to the Raman pump
power (in mW), provided there is no saturation.
But, the gain is also a function of the effective area of the fiber. Smaller fiber cores will translate
to higher power density, and therefore, to higher gain.
It is therefore a common
procedure to identify the
fiber type and determine its
effective area when installing
the Raman pump, and to
adjust it accordingly, as
shown in the adjacent figure.
Figure 99. Raman gain distribution for various commercially available fibers.
Today’s procedure might involve tweaking the Raman gain on‑site and monitoring the gain at
the receiving end; however, this requires time, effort and team synchronizations involving many
individuals. More often than not, the Raman scripts are based on the expected or assumed
fiber type. If any surprises do occur, like higher‑than‑expected loss or misidentification of the
fiber type, the entire turn‑up will shut down and be reconfigured. Raman amplification is for
long‑haul, multispan systems. The technician starts with the first span, and then moves on to the
next until he or she reaches the end, at which point the return signal is initiated. If any one span
was incorrect in the Raman gain adjustment (or the fiber‑type assumption), the entire link will
be faulty. And, finding the wrong span can be a huge challenge that is likely to delay installation
and commissioning of the link by several days, in addition to requiring several truck rolls in an
attempt to isolate the issue.
As mentioned earlier, CD testing provides good insight into the fiber type at hand. Accordingly,
CD testing for this aspect of 100 Gbit/s coherent transmission is recommended.
Potential test tools for CD issues in 100 Gbit/s systems are as follows:
• FTB‑5800 CD Analyzer for ultra‑long haul
• FTB‑5700 CD Analyzer for metro links, Raman amplification and fiber‑type information
• FTB‑5240S‑P OSA with WDM Investigator for XPM and guard‑band optimization
However, “most often” does not mean “always” due to the nature of PMD: a stable average, yet
very high instantaneous maximums.
So, while most coherent DSPs will compensate for a large amount of average PMD (as per CD,
the values of tolerance differ depending on the DSP and optics quality, etc.), some high values
will nonetheless cause network failure.
This basically means that a high local extreme or rapid transition in DGD could imply that the DSP
is failing to track and compensate. The system (DSP algorithms) will attempt to find and retrieve
DGD information in order to resume compensation, but it will not compensate while searching,
even if the instant DGD falls back within the compensation range.
The best way to reduce the probability that any of these issues occur is to use fibers with low
PMD. PMD testing is therefore paramount for successful operation of 100G systems.
Exception Case
An exception case occurs when a section offers high‑polarization scrambling, and occurs right
before a section of slightly higher PMD. High‑polarization scrambling can occur due to any of
the following:
• Aerial fiber
• Fiber under bridges
• Underground fiber alongside highways, railways, etc.
• Any situation that can cause vibration in the fiber
When this scenario occurs, the scrambling caused by the vibrating fiber can greatly influence
the state of the polarization conditions entering the high PMD section. This could cause abrupt
changes in PMD, and much higher and more frequent DGD Max than expected at random times
and wavelengths. The Maxwellian distribution referenced above does not hold true in those
scenarios. Any link that exhibits fairly high PMD (even if it is below the specified 100 Gbit/s
tolerance) and contains vibrating sections should be inspected closely using, for example,
a distributed PMD analyzer at the sections bordering the vibration.
Therefore, despite the fact that vendors claim extreme tolerance to PMD, testing for PMD remains
a required part of 100 Gbit/s testing. For obvious reasons, a recognized method for buried and
aerial fibers that is traceable to standards is preferred.
Potential test tools for PMD issues in 100 Gbit/s systems are as follows:
• FTB‑5500B PMD Analyzer for ultra‑long‑haul testing
• FTB‑5700 PMD Analyzer for metro links
• FTB‑5600 P‑OTDR for PMD mitigation
• FTB‑5240S‑P OSA with WMD Investigator for troubleshooting
The importance of OSNR is well‑known. OSNR assesses the signal quality, providing signal‑quality
indication at all wavelengths simultaneously (whereas BER only assesses one wavelength at a
time). In most cases, it can predict whether the BER is likely to pass or fail, and it is therefore of
enormous importance during commissioning for proper system turn‑up, optimization and time
savings. In troubleshooting cases, it is the perfect tool for determining the presence and cause
of high‑noise sources.
In a nutshell, better OSNR leads to lower BER, which in turn leads to higher QoS. As such,
in the very worst cases, low OSNR can be the root cause of lack of transmission, but more
commonly leads to errors in transmission, increased downtime, extra truck rolls, lower QoS and
SLA penalties.
Today’s 100G coherent detection systems are, as previously discussed, much more robust
in withstanding physical impairments. Provided that the transmitter and receiver synchronize,
transmission should be error‑free. However, it is an accepted fact that the main cause for Tx‑Rx
losing synchronization is OSNR, and therefore OSNR should be the de facto test conducted
prior to any coherent deployment.
The following two charts, which are sourced from Fabio Moliner García’s “Analysis of the OSNR
of the Links of Optical Fibre,” Universidad Autónoma de Madrid, 2009, clearly show the OSNR
and BER relationship (each curve is system‑specific, depending on dispersion, modulation format,
rate, channel count, etc.; however, all other things being equal, the OSNR‑BER relationship is clear).
Figure 104. Relationship between OSNR and BER: a higher OSNR means a lower BER.
The importance of OSNR in coherent systems is further reinforced by the fact that these systems
are not planned for metro or long‑haul applications, but rather for ultra‑long‑haul applications
spanning over several hundred or even thousands of kilometers. Each amplifier, whether it is
EDFA or Raman, increases the noise. This, combined with the fact that for a similar Tx power,
the power is divided into two phases and two polarizations that cause each symbol to have ¼
of the total power, or 6 dB less than non‑DP‑QPSK systems, meaning that the pressure on the
OSNR makes it a prime candidate for network failures.
In non‑coherent systems, and in the absence of ROADMs, the interpolation method is currently
used and generally accepted.
The IEC method described above assumes that the noise is visible between channels, and that
it is flat throughout the spectrum. These two conditions are true in most legacy networks, but
false in next‑generation networks.
The second case arises when 40G/100G is present. The spectral shape of these wavelengths
is much larger than that of a 10G wavelength (typically two and a half to four times wider).
This prevents the spectrum analyzer from seeing the noise level between two large signals.
Again, in the example below, the red arrow represents the point at which the IEC method would
measure the noise, whereas the green arrow represents a more probable noise level.
Figure 107. Noise level of 40G channels. Red arrow: noise level using the IEC method. Green arrow: real noise level.
The following chart represents a true operating transmission line, and is divided into three
sections to highlight the previously explained details.
Figure 108. Spectrum with 10G channels in a ROADM network and 40G channels.
The first section to the left in the chart above (i.e., from 1527 nm to 1545 nm) represents ASE
noise. Normally, ASE noise is somewhat flat and uniform. The example in the chart has gone
through a certain amount of ROADMs, and therefore the shapes of the different residual filters
can be observed. This proves that in such systems, one of the assumptions of the IEC OSNR
measurement method, i.e., that noise is flat and uniform, is false.
The second section (i.e., from 1545 nm to 1553 nm) displays narrow 10G transmission in a
ROADM mesh network, which will create a false sense of security. This example clearly indicates
the level of noise that the IEC method would measure versus the true noise experienced by the
channel. In this case, the error is about 3 dB, which in many cases is a toggle between system
acceptance and failure. This system could experience BER despite the fact that all the parameters
seem OK.
Finally, according to the information displayed in the third section (i.e., from 1553 nm to 1562 nm),
the user would get an inaccurate representation of the problem if using a traditional IEC‑based
OSA, as demonstrated by the different noise levels highlighted. The network manager could
spend hours trying to understand, from a network point of view, why the OSNR is so low, when in
fact it could be just fine. So, a single link can contain wavelengths that are over or underestimated
based on the testing method.
It is worthy to note that most vendor management systems (VMS) use this OSNR definition to
compute the OSNR, and to balance and attempt to optimize the system, which could lead to
significant and costly mistakes.
• OSNR = 10log(R)
• s(λ): time‑averaged signal spectral power density, not including ASE, expressed in W/nm
• ρ(λ): ASE spectral power density, independent of polarization, expressed in W/nm
• Br: reference bandwidth expressed in nm (usually 0.1 nm)
• λ1 to λ2: signal spectral range
As previously mentioned, historically, OSNR has been based only on noise coming from the ASE.
But, only measuring OSNR in this manner could lead to serious errors. For example, boosting
the power to increase the ASE‑based OSNR may actually lead to worse BER.
The following figure shows that once a certain level of OSNR is reached (21 dB to 22 dB in this
example), the spectral efficiency starts to decrease again, creating additional BER.
The pink line in Figure 110 (above) represents the pure ASE non‑linearity trace. The blue line
represents WDM non‑linearities, and the green line represents single ASE plus optical filtering.
The red line represents the resulting trace from a real signal combining the different effects of
ASE noise and non‑linearity impairments.
84 EXFO—100G/400G Testing Guide—Line Side
As indicated, at a certain point the transmission quality will be limited by non‑linearity. As such, if
the BER is insufficient, it will be difficult to determine whether the cause of the impairment is ASE
noise or non‑linearity. Most optical spectrum analyzers deliver OSNR based on ASE noise only,
which is part of the actual definition, but do not account for non‑linear noise and OSNR contribution.
As discussed in the section devoted to chromatic dispersion, NLEs can now be monitored using
the WDM Investigator dashboard while the OSA delivers a direct ASE noise measurement
independent of other noise sources. In addition, noise is defined two‑fold (whenever
distinguishable and independently measurable), as follows: ASE noise and extended noise (taking
into account ASE and non‑linear noise). The BER estimate should be based on this extended
noise and extended OSNR.
Figure 111. WDM Investigator reading showing a non‑linear depolarization contributing to extended noise (OSNRe).
The next step involves in‑band OSNR applied to Pol‑Mux signals.The basic setup of an in‑band
OSA, as shown in Figure 112, consists of using a polarization controller followed by a polarization
beam splitter. In the case of non Pol‑Mux signals, given that the signal is polarized and that the
noise is unpolarized, adjustments to the polarization controller will change the proportion of signal
split into each of the two branches, SOP‑1 and SOP‑2. To compute the in‑band OSNR, complex
algorithms can then be applied to the OSA traces found in branches one and two.
SOP‑2
SOP‑1
Figure 112. In‑band OSNR setup.
There are currently two standards applicable for Pol‑Mux OSNR: the IEC 61282‑12 mentioned
previously, as well as the China Communications Standards Association (CCSA) YD/T
2147‑2010.
The Chinese CCSA YD/T 2147‑2010 standard recommends calculating Pol‑Mux OSNR as follows:
Without going into too much detail, both of these standards are based on the channel turn‑off
method, which consists of measuring the signal plus noise power with the channel turned on,
and then measuring the noise power with the channel turned off.
EXFO’s Pol‑Mux OSNR measurement tool, called the commissioning assistant, is a software
option available in EXFO’s FTBx‑5245/5255 OSA that complies with both the IEC 61282‑12
standard and the CCSA standard.
To utilize the commissioning assistant, the user must first take a measurement at the receiver
with all of the channels turned on (as shown in Figure 113), and then acquire a series
of traces that have each been taken with one channel turned off (as shown in Figure 114).
The commissioning assistant therefore requires a total of m+1 traces, where m represents the
number of channels. The commissioning assistant then performs the Pol‑Mux OSNR calculations
via a user‑friendly wizard.
Figure 113. Trace with all channels on, ready for Pol‑Mux OSNR calculation
with the commissioning assistant.
Figures 115 to 117 show the most common combinations of reference and active traces.
Figure 115 describes a case where a reference has been acquired at time t0, commonly at
turn-up, and an active trace is taken at a later time (t), say 6 months later. On the other hand,
figure 116 displays a case where both the reference trace and active trace are taken at roughly
the same time (t), in which the reference Pol-Mux OSNR values are known. Figure 117 shows
a case that can be very useful if no prior information about the network is available. In this case,
the reference acquired at the transmitter Tx can be assumed to be noise free, because almost all
the noise comes from the optical amplifiers, which the signal has not yet travelled through. The
OSNR value can then be manually entered as 35 or 40 dB in the reference trace.
Tx Rx
DEMUX
MUX
Figure 115.
Tx Rx
DEMUX
MUX
Figure 116.
Tx Rx
DEMUX
MUX
EDFA EDFA EDFA EDFA
Figure 117.
The basic measurement principle consists of subtracting a noise-free reference from the active
trace, which contains signal + noise, in order to obtain the noise level (figure 118).
In general, a normalization step is needed when the power of the active and the reference traces
are different.
A noise-free reference is only readily available at the transmitter location. However, one can
compute a noise-free reference from any location provided that the OSNR or noise level at this
location is known. To do so:
• The first step consists of computing the noise level of the reference from the OSNR value
• The second step consists of subtracting the reference noise level from the reference
spectrum
• A noise-free reference is then obtained (figure 119)
All these steps are carried out using a user-friendly wizard in EXFO OSA— the in-service Pol-
Mux assistant.
0
Bit Error Rate Vs Error Vector Magnitude(rms)
10
BPSK
4−QAM
−1
10 16−QAM
64−QAM
BPSK(Theoritical)
−2
10 4−QAM(Theoritical)
16−QAM(Theoritical)
Bit Error rate
64−QAM(Theoritical)
−3
10
−4
10
−5
10
−6
10
−30 −25 −20 −15 −10 −5 0
Error Vector Magnitude (rms) in dB
Figure 120. BER vs. EVM relationship for different modulation formats.
According to one NTT Worldwide Telecommunications survey, more than 70% of optical network
faults arise from fiber endface inspection. A dirty or damaged endface of an optical fiber connector
leads to increased insertion loss, return loss and even connector burnout in extreme cases.
The hybrid method combines dry and wet cleaning. First, the connector endfaces need to be
cleaned with solvent, after which any residual solvent must be wiped off with a wipe or cotton
swab (depending on the type of connector being cleaned). The inspection clean inspection
clean (ICIC) method is recommended in practical operation. The diagram on the next page
compares optical fiber endfaces before and after cleaning. In terms of optical fiber endface
inspection, related international standards (such as those of the IEC and IPC) have very detailed
requirements; meanwhile, operators can also define their own standard for optical fiber endface
inspection according to their respective situations.
Before Cleaning After Cleaning
Using minimal configuration, it quickly validates the soundness of an optical device and it includes
a sequence of basic sub‑tests, where parameters and thresholds are automatically set or are user
configurable.
It supports a wide range of optical modules: QSFP28, QSFP+, CFP, CFP2, CFP4, XFP, SFP+, SFP.
1. Power Monitoring
3. Temperature Monitoring
4. Optical TX Power Level Range Test (each lane is individually tested for multi‑lane devices)
5. Optical RX Signal Presence & Level Range Test (each lane is individually tested for multi‑
lane devices)
6. Stress BERT
For more details about this tool, please refer to Section 2.3.
Figure 123. TTI SAPI/DAPI/operator information test using the FTB‑88100NGE Power Blazer.
Test procedure:
a) Connect test configuration as per the above diagram. All OTU4 services must be
concatenated respectively. The services should be activated normally.
b) Configure the client signal as 100 Gbit/s, then start the long‑term BER test.
c) Record the test results after 24 hours.
The following sections cover EtherSAM or ITU‑T Y.1564, the ITU‑T standard for turning up,
installing and troubleshooting Ethernet‑based services. EtherSAM is the only standard test
methodology that allows for complete validation of Ethernet service‑level agreements (SLAs) in
a single, significantly faster test with the highest level of accuracy.
To assure QoS, providers need to properly configure their networks to define how the traffic
inside will be prioritized. This is accomplished by assigning different levels of priority to each type
of service and accurately configuring network prioritization algorithms. QoS enforcement refers
to the method used to differentiate the traffic of various services via specific fields in the frames,
thus prioritizing frames for certain services over other frames. These fields make it possible for a
network element to service and discriminate between high‑ and low‑priority traffic.
Best‑Effort Data
Key Performance Indicators Real‑Time Data High‑Priority Data
(Internet Access)
CIR (Mbit/s) (green traffic) 2.5 5 10
EIR (Mbit/s) (yellow traffic) 5 0 5
Frame delay (ms) <30 <5 5‑15
Frame delay variation (ms) n/a <1 n/a
Frame loss (%) <0.05 <0.001 <0.05
VLAN 300 100 200
Figure 126. Key performance indicators (KPIs) for various traffic types.
Customer traffic is classified into three traffic classes, each of which is assigned a specific
color in Figure 127 (below): green for committed traffic, yellow for excess traffic and red for
discarded traffic.
• Bandwidth: Bandwidth refers to the maximum amount of data that can be forwarded.
This measurement is a ratio of the total amount of traffic forwarded during a measurement
window of one second. Bandwidth can either be “committed” or “excess,” with different
performance guarantees. Bandwidth must be controlled, because multiple services typically
share a link. Therefore, each service must be limited to avoid affecting another service.
Generating traffic over the bandwidth limit usually leads to frame buffering, congestion, and
frame loss or service outages.
• Frame Delay (Latency): Frame delay, or latency, is a measurement of the time delay
between the transmission and the reception of a packet. Typically, this is a round‑trip
measurement, meaning that it simultaneously calculates the near‑end to far‑end directions,
as well as the far‑end to near‑end directions. This measurement is critical for voice
applications, because too much latency can affect call quality, leading to the perception of
echoes, incoherent conversations, or even dropped calls.
• Frame Loss: Frame loss can occur for numerous reasons, including transmission errors
and network congestion. Errors due to a physical phenomenon can occur during frame
transmission, resulting in frames being discarded by network devices such as switches
and routers based on the frame‑check‑sequence field comparison. Network congestion
also causes frames to be discarded, because network devices must drop frames to avoid
saturating a link in congestion conditions.
• Frame Delay Variation (Packet Jitter): Frame delay variation, or packet jitter, refers to the
variability in arrival time between packet deliveries. As packets travel through a network,
they are often queued and sent in bursts to the next hop. Random prioritization may occur,
resulting in packet transmission at random rates. Packets are therefore received at irregular
intervals. This jitter translates into stress on the receiving buffers of the end nodes, which
may result in buffers becoming overused, or underused in the event of large swings of jitter.
Real‑time applications such as voice and video are especially sensitive to packet jitter.
Buffers are designed to store a certain quantity of video or voice packets, which are then
processed at regular intervals to provide a smooth and error‑free transmission to the end
user. Too much jitter will affect the quality of experience (QoE), because packets arriving at
a fast rate will cause the buffers to overfill, resulting in packet loss. Packets arriving at a slow
rate will cause buffers to empty, leading to still images or sound.
EtherSAM enables complete validation of all SLA parameters within a single test, thus ensuring
optimized QoS. Contrary to other methodologies, it supports new multiservice offerings. In fact,
EtherSAM can simulate all the types of services that will run on the network, and simultaneously
qualify all key SLA parameters for each of these services. In addition, it validates the QoS
mechanisms provisioned in the network to prioritize the different service types, resulting in more
accurate validation, and much faster deployment and troubleshooting. Moreover, EtherSAM offers
additional capabilities, such as bidirectional measurements.
EtherSAM (ITU‑T Y.1564) is based on the principle that the majority of service issues are found
in two distinct categories: a) in the configuration of the network elements that carry the service,
or b) in the performance of the network during high‑load conditions, when multiple services
cause congestion.
Service Configuration
Forwarding devices, such as switches, routers, bridges and network interface units, are the
foundation of any network, because they interconnect segments. These forwarding devices must
be properly configured to ensure that traffic is adequately groomed and forwarded in accordance
with its SLA.
If a service is not correctly configured on a single device within the end‑to‑end path, network
performance can be greatly affected. This may lead to service outages and network‑wide issues,
such as congestion and link failures. Therefore, a very important part of the testing effort is
ensuring that devices are properly configured and able to handle the network traffic as intended.
Service Performance
Service performance refers to the ability of the network to carry multiple services at their
maximum guaranteed rate without any degradation in performance; i.e., KPIs must remain within
an acceptable range.
As network devices come under load, they must make quality decisions. They must prioritize
one traffic flow over another in order to meet the KPIs of each traffic class. This is necessary,
because as the amount of traffic flow increases, so does the likelihood of performance failures.
Service performance assessment must be conducted over a medium‑ to long‑term period, because
problems typically occur in the long term and will probably not be seen with short‑term testing.
The combination of these two tests provides all the critical results in a simple and complete test
methodology. The service configuration test quickly identifies configuration faults by focusing
on each service and how it is handled by the network elements along the paths. The service
performance test focuses on the network’s capacity to handle and guarantee all services
simultaneously. Once both phases have been successfully validated, the circuit is ready to be
activated and placed into service.
This emulates real‑life network traffic, and can uncover network‑equipment configuration issues
that could go undetected with non‑simultaneous bidirectional testing. Furthermore, performing
simultaneous bidirectional testing significantly reduces costs by decreasing test time by 50%.
Carriers and service providers face the constant challenge of ensuring the proper delivery of
services to customers. Ethernet services need to be delivered to customers in compressed time
frames, while proving to be more reliable than ever. EtherSAM bridges the gap between service
validation and performance assessment by providing an intuitive and easy approach for confident
control and management of networks while reducing OPEX and growing revenues. EtherSAM
is the only standard test methodology that allows for complete validation of SLAs in a single,
significantly faster test, while offering the highest level of accuracy.
iSAM does this by making service turn‑ups what they should be: simple and quick. Inspired by
the feedback and field‑testing knowledge of customers around the world, iSAM was designed to
simplify field technicians’ daily tasks by focusing on the most common Ethernet service turn‑up
configurations used in the industry today, from 10M all the way up to 100G.
Imagine that field technicians could set up an Ethernet service test within a single page,
automatically connect to their remote unit, complete testing, and then have the test report
automatically uploaded to a cloud‑based storage location. iSAM makes this possible.
In today’s test and measurement (T&M) market, field technicians often waste valuable time going
through a multitude of test setup pages. Thanks to its innovative new approach, iSAM can be
configured entirely within a single page for maximum efficiency.
Due to the shift in customer usage patterns and the demand for bandwidth‑intensive applications,
service providers must also adapt their testing requirements to meet customer expectations.
Fifteen years ago, service providers rarely talked about testing or validating the TCP layer.
However, today they need to go beyond traditional RFC 2544 testing, and even ITU‑T Y.1564
service activation testing.
With the simplicity that iSAM brings to the table, service providers also have the ability to easily
validate their services right up to TCP layer 4. Simply enabling the RFC 6349 test in the iSAM
setup page will allow the TCP Throughput validation test to be performed automatically after
layers 2 and 3 have been tested. This will ensure validation of the base layers (Ethernet and IP)
along with the TCP layer, thereby delivering a true end‑to‑end service activation test.
The RFC 6349 test will automatically determine the path maximum transmission unit (MTU)
to avoid packet fragmentation. The test will also measure the baseline round‑trip time (RTT),
which constitutes the minimum round‑trip time required for a TCP packet to be sent out
and acknowledged by the receiver. From the baseline RTT and the user‑entered committed
information rate (CIR), the bandwidth delay product (BDP), or optimum window size, will be
calculated. The BDP (optimum window size) will be used to perform the TCP throughput test.
Field technicians will then be able to quickly determine whether the circuit under test is operating
at the expected level.
Designed according to the same principles as EtherSAM, iSAM also integrates the concept of
two subtests: the configuration test and the performance test. The configuration test verifies
the performance according to per‑service‑basis criteria. This test verifies that the performance
metrics defined by the MEF, such as frame loss ratio (FLR), frame delay (FD) and inter‑frame delay
variation (IFDV), are met at the CIR for each individual Ethernet service. The performance test,
on the other hand, verifies that the performance metrics are met while all services are operating
simultaneously at the combined CIR.
Over the years, many service providers have searched for guidance regarding the proper
performance metric settings needed to test Ethernet services. For instance, should the FLR be
below 0.5% for a short‑span service such as the metro service? Should the FD be less than
10 ms, 20 ms, or 30 ms? There has been quite a bit of confusion concerning the right metrics
needed based on the type of Ethernet service provided. iSAM clears up this confusion by
providing MEF performance profiles to field technicians. With iSAM, service providers and field
technicians can now easily test according to these guidelines based on the Ethernet service
being delivered.
MEF 23.1 specifies three performance objectives and four performance tiers (PTs).
The performance objective relates to the class of service (CoS) being delivered, whereas the
performance tier relates to the span of the Ethernet service.
Performance Objectives
H (High) Stringent performance, such as timing distribution
M (Medium) Real‑time applications, e.g., voice‑over‑Internet protocol
L (Low) Best‑effort applications (web traffic)
Because iSAM includes an extensive list of Ethernet services, service providers no longer have
to play a guessing game when turning up and validating Ethernet services. Field technicians can
easily select the desired performance criteria based on the Ethernet service being delivered to
the end customer.
Performance‑Criteria Customization
What happens when the MEF predefined performance criteria do not meet the user’s needs?
With iSAM, users also have the ability to modify the predefined MEF performance criteria or create
their own user‑defined performance‑criteria profiles. Some service providers may even have their
own Ethernet service offering (e.g., Gold, Silver and Bronze) with associated performance criteria
for these services. With iSAM, custom performance‑criteria profiles can easily be created.
These custom performance‑criteria profiles are created using master templates that are
accessible on the platform (not within the test application), and can be restricted to managers
and/or supervisors. Also, if a user has modified an MEF performance‑criteria template, an asterisk
will appear near the performance‑criteria name to indicate that some modifications to the metrics
have been carried out.
The difficulty of the associated tasks is compounded by the pressure of limited investigation time
and the risk of customers being affected.
One of the options available to network engineers is to capture and decode traffic on the affected
circuit. Decoding traffic usually involves analyzing the header traffic to identify any issues, such
as modifications or incorrect content.
The filter engine is based on basic and advanced filtering capabilities. The filter can be configured
based on different logical operands (AND, OR, NOT). The figure below shows the filtering flow,
where the Ethernet packet can be passed through and/or selected for capture and analysis.
7 1 6 bytes 6 bytes 2 4
Preamble SD MAC Dst MAC Src Type Information Pad FCS
Address Address
7 1 6 bytes 6 bytes 2 4
Capture Engine
7 1 6 bytes 6 bytes 2 4
MAC Dst MAC Src
Preamble SD Address Address Type Information Pad FCS
Through Preamble SD MAC Dst
Address
MAC Src
Address Type Information Pad FCS
7 1 6 bytes 6 bytes 2 4
Filter 7 1
Address
6 bytes
Address
6 bytes 2 4
Preamble SD MAC Dst MAC Src Type Information Pad FCS MAC Dst MAC Src
Address Address Preamble SD Address Address Type Information Pad FCS
7 1 6 bytes
MAC Dst
6 bytes
MAC Src
2 4 Selected
Preamble SD Address Address Type Information Pad FCS
7 1 6 bytes 6 bytes 2 4
Preamble SD MAC Dst MAC Src Type Information Pad FCS
Address Address
7 1 6 bytes 6 bytes 2 4
Preamble SD MAC Dst MAC Src Type Information Pad FCS
Address Address
In most captures, the payload information is typically proprietary information that cannot be
decoded by the test equipment. As such, network engineers usually focus on the header
information, which is decoded and used for more in‑depth troubleshooting. Again, this eliminates
the capture of unnecessary data that would otherwise consume memory without providing any
additional useful information.
Capture Trigger
A very common issue with typical capture tools is that capture starts as soon as the capture tool is
enabled. However, the event of interest may occur later, allowing previously captured traffic to fill
the memory buffer without providing any useful information. In some cases, the testing opportunity
can be completely missed due to the high amount of captured data and the short event window.
EXFO’s Ethernet capture tool solves this issue by including a set of triggering capabilities that
allow customers to fine‑tune and specify when the capture process should start. This powerful
capability simplifies the troubleshooting process by filling the memory only when the event of
interest is detected. The memory and the troubleshooting time are therefore efficiently used,
resulting in meaningful capture data that yields more important information.
Triggering Position
The triggering position is used to determine the position of the triggered frame within the captured
data, solving a common problem with traditional capture tools, whereby the event of interest is
often located within the capture data.
A typical use for the triggering position is to perform pre‑ and post‑analysis. In network
troubleshooting, it is very important to understand the events that led to the failure and to view
the events that followed the failure. These two critical phases provide a wealth of information
about the failure, what caused it, and how the network reacted to it.
The triggering‑position capabilities allow the user to specify where the trigger event will be
located in the capture, making it possible to select the frames to be captured depending on their
position relative to the trigger event. Traditional capture tools do not offer the ability to perform
a mid‑trigger or pre‑trigger; instead, they only provide post‑trigger capabilities. As such, users
are left to manually search within the captured sequence to identify the event and perform the
analysis. This, in combination with the lack of a trigger mechanism, means that it is possible
to completely miss the event of interest when using traditional capture tools, resulting in an
inefficient capture process.
Conclusion
With 100G moving toward mass deployment, the Ethernet capture and decode capability offered
by the FTB‑88100NGE, FTBx‑88200NGE and FTB‑890/890NGE field testers allow network
engineers to quickly pinpoint issues in the field and speed up the troubleshooting process to
ensure quick service recovery.
Ethernet service OAM (SOAM), on the other hand, provides management solutions to guarantee
and measure end‑to‑end performance. SOAM enables end‑to‑end SLAs for standardized
Ethernet services, in‑service SLA verification, as well as network monitoring and troubleshooting
from the central office. Also, SOAM protocols support two sets of functions, the first of which
is connectivity fault management (CFM), which detects, verifies and isolates connectivity
failure as defined in ITU Y.1731, IEEE 802.1ag and MEF 30.1. This is performed end‑to‑end,
although some functions can isolate faults in segments. The second set of supported functions is
performance monitoring (PM), which provides the required capability for performance monitoring
as defined in ITU Y.1731 and MEF 35. This is performed from end to end. Other OAM protocols
include IEEE 802.1ab for link‑layer discovery, IEEE 802.3ah for Ethernet in the first mile, and
ITU‑T G.8113.1 for multiprotocol label switching−transport profile (MPLS‑TP) OAM. The Metro
Ethernet Forum has defined additional service OAM requirements, namely MEF 17.
Interface
User Network
Equipment
Edge
Equipment
Edge
Interface
User Network
Equipment
Customer
CFM CFM
Performance Monitoring
• Maintenance entity group end point (MEP): The boundary points of a maintenance
domain, they can initiate and terminate OAM frames. End‑to‑end tests are initiated and
terminated by MEPs.
• Maintenance entity group intermediate point (MIP): The intermediate points in a
maintenance domain, they do not initiate OAM frames, but can respond to some OAM
frames (loopback and link trace) in order to isolate faults.
• Maintenance entity (ME): This entity requires management and defines a relationship
between two maintenance entity group endpoints.
• Maintenance entity group (MEG): A group of MEs that are in the same administrative
boundary, with the same MEG level, and belonging to the same point‑to‑point or multipoint
Ethernet connection.
• Maintenance association (MA): A set of MEPs established to verify the integrity of a single
service instance.
• IEEE 802.1ag: 802.1ag focuses on the end‑to‑end connectivity and continuity of nodes in
an Ethernet network. This is why it is referred to as connectivity fault management, or CFM.
Because it applies to bridges and bridge applications, it specifies a lot of multicast packets in
addition to unicast packets, and also handles both multipoint and point‑to‑point connections.
802.1ag has three main functions: continuity check messages (CCM), loopback messages
(LBMs) and loopback responses (LBRs), and link trace messages (LTMs) and link trace
responses (LTRs).
• CCM: A continuity check message is an OAM protocol data unit (OAMPDU) that provides
service monitoring from one endpoint to another (MEP to MEP). The CCMs exchanged
between MEPs are analogous to a “heartbeat” message, and can be configured to be sent
at one of seven standard intervals: 3.3 ms, 10 ms, 100 ms, 1 s, 10 s, 1 min, and 10 min.
CCMs can be either multicast or unicast, although the use of multicast packets, which run
continuously until turned off, is preferred. When the reception of CCM messages at a node
is lost, this constitutes a loss of connectivity to that node. The reception of CCM messages
from an unknown node represents a possible misconfiguration of nodes.
• LTM/LTR: LTMs and LTRs are used by an MEP to verify the complete link trace to a peer
MEP. Each MIP and the peer MEP will respond to an LTM. Once a network fault has been
confirmed by the service provider, the link trace feature can be used to isolate its specific
location. This feature traces the service from one MEP to another MEP or MIP using its
MAC address, and to all the MIPs along the MA.
• ITU‑T Y.1731: The ITU‑T Y.1731 protocol is used mainly for fault management and
performance monitoring, defining performance monitoring measurements (i.e., frame loss
ratio, frame delay, frame delay variation, service availability) in order to assist with SLA
assurance and capacity planning. This protocol applies to both multipoint and point‑to‑point
connections, and relies on the 802.1ag protocol for transport, making it a type of extension
to 802.1ag. The ITU‑T Y.1731 protocol measures the following performance parameters:
frame loss, delay, delay variation and service availability. Its main features are alarm
indication signals (ETH‑AIS), remote defect indication (ETH‑RDI), locked signal (ETH‑LCK),
test signal (ETH‑Test), performance monitoring (ETH‑PM), frame loss measurement
(ETH‑LM), frame delay measurement (ETH‑DM) and client signal fail (ETH‑CSF).
CE Operator A Bridges Operator B Bridges CE
1 2 3 4 5 6 7 8 9
Cala
ETH
• REF: ITU-T Y.1731 Fig. II.1
• ETH‑AIS: This message is sent to the far end when the near end detects an alarm via the
CCM packets. The AIS frame is transmitted periodically until the fault condition is resolved.
• ETH‑RDI: This indication is used for fault notification. If an MEP is defective, it will transmit
an RDI to its peer MEPs to inform them that a defect condition has been encountered.
• ETH‑LCK: This signal is used to communicate an administrative lock to the far end, resulting
in an interruption in data traffic. The signal tells the far end that the near end is present,
but unavailable for use. LCK frames are also transmitted periodically until the administrator
clears the lock.
• ETH‑PM: This functionality is used to monitor the performance of traffic from point to point
or from end to end on a given domain.
• ETH‑LM: This feature is used by an MEP to measure the frame loss with the peer MEP in
both directions from a single endpoint.
• ETH‑DM: This functionality is used by an MEP to measure the roundtrip delay with the peer
MEP. This is also used to measure the delay as well as the delay variation.
• ETH‑CSF: This feature is used for fault notification. It is used by an MEP to propagate to a
peer MEP the detection of a failure or event in an Ethernet client signal. It is used when the
client does not support fault detection mechanisms.
• IEEE 802.3ah: This standard is also referred to as Ethernet for the First Mile (EFM) OAM.
While 802.1ag specifies verification of end‑to‑end connectivity across multiple domains and
hops, 802.3ah specifies verification of point‑to‑point connectivity across only one hop.
It uses the following mechanisms: discovery, loopback, polling of management information base
(MIB) variables, remote failing indication and link monitoring.
• Discovery: This is the first phase. This is where the devices in the network are identified
along with their OAM capabilities. If the discovery fails for any reason, all other OAM activity
is aborted until the discovery can be re‑established.
• Loopback: An OAM entity can put its remote peer into loopback mode using loopback
control payload. This helps the administrator ensure link quality during installation or
troubleshooting.
• Polling of MIB Variables: IEEE 802.3ah provides read‑only access to remote MIB, but this
is limited to specific MIB branches and leaves. This feature is based on the premise that the
administrator can also retrieve/reset MIB variables at the far end. These variables are one of
the primary sources of OAM information available to the system administrator.
• Remote Failure Indication: This mechanism enables the OAM entity to convey the
degradation of an Ethernet link to its peers via specific tags in OAM payload.
• Ethernet link trace: This functionality, which is used for fault isolation, allows an MEP to
verify the complete link trace to a peer MEP. Each MIP and its peer will respond to the link
trace message.
Network equipment supporting OAM features is being massively deployed in networks carrying
Ethernet services due to the essential functions it provides in those networks. As such, it is
imperative to ensure that this equipment is properly configured and delivering on the promised
features and functionality. Testing OAM services prior to deployment will therefore help service
providers and carriers save time and money.
• Fault Management: Depending on the OAM standard being used, several types of tests
can be performed to allow the detection, verification, location and notification of various
defect conditions.
• Continuity Check: This function is used to confirm the presence of an MEP or test
equipment on a network, and to verify the presence of peer MEPs. During this test,
transmitted frames are received either by a peer MEP using a unicast destination address,
or by all the MEPs in the MEG using a multicast destination address.
• Loopback Test: During this test, the tester transmits an LBM payload with a specific
sequence number. The peer MEP responds to the LBM with an LBR payload. The test
validates each received LBR and reports any invalid LBR, invalid payload and LBR timeout.
• Link Trace Test: This test verifies that the complete link trace reaches the peer MEP.
The tester will send LTM messages and receive LTR messages.
• Test Function: During this phase, the tester generates frames with a specific test pattern
and sequence number in order to verify the integrity of the signal received by the peer MEP.
This test requires two testers, one at each end.
• RDI Test: During this phase, the tester generates a remote defect indicator (RDI) to simulate
a defect and validate the reaction and behavior of the peer MEP.
• CSF Test: During this phase, the tester generates and detects a client signal fail (CSF) in
order to validate the reception and behavior of the peer MEP.
• Frame Delay Test: In this test, which is a critical OAM metric, the tester measures the
roundtrip delay to the peer MEP. In order to simulate real‑world conditions, different frame
sizes should be used to validate the frame delay.
• Frame Loss: This test checks the bidirectional frame loss occurring with a peer MEP from
a single endpoint, and should be done over as long a period of time as possible in order
to obtain a good idea of how the network will behave. To simulate real‑world conditions,
different frame sizes should be used to validate the frame loss.
• Synthetic Loss: This test uses synthetic frames to check the bidirectional frame loss
occurring with a peer MEP from a single endpoint. This test should also be done using
different frame sizes.
Conclusion
As multiservice networks get more complex, new technologies will necessarily emerge.
Even though the concept of operating, administrating and maintaining networks has existed since
the early days of synchronous optical networks and synchronous digital hierarchy (SONET/SDH),
Ethernet OAM is continually evolving, with new standards being developed as stakeholders are
just beginning to understand existing ones.
OAM plays an important role in networks, and as such, these networks must be properly
configured to support all the features it offers. However, this can only be achieved by thoroughly
testing all the applicable parameters mentioned in this document. From a practical standpoint,
because the latest technology is not always fully understood by all users, the most efficient
approach available to network operators is to use testers equipped with all the required OAM
features and metrics.
POTN is built to the OTN G.709 standard based on ITU‑T and is also featured by the packet
network. The conventional G.709 is based on circuit‑switching technology, which is advantageous
in its ability to guarantee zero packet loss, low delay, small packet delay variation and superior
synchronous clock performance. The conventional packet network improves transport efficiency
through statistical multiplexing and features a flexible service, but it cannot offer high clock
performance as per the OTN network, e.g., high‑quality transport. This is where the POTN
network comes into play: a single device integrates circuit‑switching characteristics inherent
to the OTN network while incorporating the highly-efficient statistical multiplexing feature of
packet switching.
POTN networks adopt MPLS‑TP technology, which is a simplified version of MPLS for transport
networks from which some of the MPLS functions have been removed, including penultimate hop
popping (PHP), label‑switched path (LSP) merge, and equal cost multipath (ECMP). MPLS‑TP
does not require MPLS control‑plane capabilities, and enables the management plane to set
up LSPs manually.
The functions of OAM for MPLS‑TP networks are intended to reduce the operational complexity
associated with network performance monitoring and management, fault management and
protection switching. One of the goals of MPLS‑TP OAM is to provide the tools needed to monitor
and manage the network with the same attributes offered by legacy transport technologies.
Two important components of the OAM mechanisms are generic associated channel (G‑Ach)
and generic alert label (GAL), which allow an operator to send any type of control traffic into
a pseudowire (PW) or label‑switched path (LSP). G‑ACh is used in both PWs and MPLS‑TP
LSPs, whereas GAL is used in MPLS‑TP LSPs to flag G‑Ach.
The OAM functionality of MPLS‑TP used in fault management consists primarily of two kinds of
management, namely proactive FM OAM and on‑demand FM OAM functions.
Traditional TDM networks were devoid of such parameters and only provided guaranteed
throughput. Inefficient and complex, they focused only on plain transmission of data. There was
no recognition of traffic type, and all traffic was serviced with the same level of service. With the
deployment of packet‑based networks, new applications must now be serviced by the networks;
however, these different applications have different performance requirements. Voice over TDM
makes very different assumptions than over‑the‑top video, broadcast TV, the Web or e‑mail. To
allow different applications to share a single backhaul circuit, the circuit must support more than
one class of service (CoS), each with its own QoS requirements. Packet‑based technologies,
on the other hand, provide differentiation mechanisms that allow the network to prioritize certain
transmissions over others and provide some transmission guarantees.
POTN retains the IP/MPLS QoS systems such as IP differentiated services (DiffServ) and MPLS
labeling, allowing the network to carry voice, video and data packets. The network is then able
to classify packets according to their priority, and service them to ensure that they meet their
performance requirements.
The Ethernet QoS test in the POTN network includes the following items: service flow classification
and priority mapping ability test; flow‑classification‑based access control list (ACL) capability test,
access rate control test; access rate control granularity, color sensitivity functional verification,
connection admission control (CAC) scheme, and client service priority remapping ability.
Test method: A typical network configuration is shown above. Create an LSP and a PW between
NE1 and NE3. Multiplex PW into LSP. Create flow classification criteria, map them into PW
and then verify the following flow classification criteria:
a. Ports f. DiffServ code point (DSCP)
b. VLAN g. IP TOS
c. VLAN PRI h. TCP/IP port number (optional)
d. Source or destination MAC address i. Any combination of the above
e. Source or destination IP address
EXFO’s SyncWatch‑110 device can also employ any external standard provided by the client,
thereby making full use of the existing synchronizing infrastructure and ensuring that the
measurement results adopt the same source as the local standard. The SyncWatch‑110 unit
provides real‑time time‑interval‑error (TIE) measurement, and enables calculation and display
of maximum‑time‑interval‑error (MTIE) and time deviation (TDEV) measurements at any point.
Test result: Figure 135 below shows the TIE test result, and Figure
136 shows the MTIE test result.
Figure 142. TIE test result. Figure 143. MTIE test result.
For the IEEE 1588v2 PTP test, the main concerns are delay assymmetry and delay variation.
The images below represent typical examples of delay symmetry. In the graph on the left‑hand
side, the master‑to‑slave delay is much higher than the slave‑to‑master delay, and therefore the
result is a fail. In the middle graph, one‑way delay is seriously degraded, which also leads to a
fail result. Ideal results are obtained when delays for each direction are symmetric.
The POTN provides the capability to handle both types of traffic through the implementation of
pseudo wire edge‑to‑edge emulation (PWE3) forwarding. This encapsulation method has been
specifically designed to segment and encapsulate TDM and packetized transmission onto packet
networks, allowing both types of traffic to be transparently transported on the POTN network,
while ensuring independence between the transport encapsulation and the encapsulated data.
Supporting CFP4/QSFP+/QSFP28
ISBN 978-1-55342-111-5
Legal Deposit—National Library of Canada 2017
Legal Deposit—National Library of Quebec 2017
©2018 EXFO inc. All right reserved. Printed in Canada 18/03 20180146V6 SAP1072302