Optical Interconnect Solution for Cloud Data Center

The cloud data centers are the infrastructure of cloud computing networks. The continuous penetration of cloud computing business has stimulated the construction of super data centers. The cloud infrastructure is mainly composed of switches and servers. Fiber optic cables and optical transceivers, or active optical cables and direct attach cables are used for the connection. The large scale of cloud data centers will greatly increase the usage of optical transceivers and the requirements to transmission distance, which will increase the usage ratio of single mode optical transceivers.

 

In the cloud data centers, the explosive growth of the traffic is driving the data rate of optical transceivers to escalate and accelerate. It has taken 5 years from 10G to 40G, then 4 years from 40G to 100G, and may likely take only 3 years from 100G to 400G. All the export data in the future data centers need to go through the internal mass operation (especially the rising internal and export flow of AI, VR/AR, UHD video, and so on). The flow of the east-west direction in the data center is turbulent, and the flat data center architecture makes the 100G optical transceiver market continue to grow at a high speed.

 

According to the report of third party institution, the number of global super large data centers will be over 500 by the end of 2019. Then 83% of the public cloud servers and 86% of the public cloud load will be loaded in the super data centers, the ratio of deployment of super data center servers will rise from 21% to 47%, the ratio of processing capacity will be increased from 39% to 68%, and the ratio of traffic will be increased from 34% to 53%.

 

 

The direction of the main data flow of traditional 3-tier data center is from top to bottom or from south to north, while the direction of the main data flow of flat spine-leaf data center is from east to west.

 

Here is a data center optical interconnect application case of optical transceivers and AOC. The network architecture of a Cloud Data Center is pided into Spine Core, Edge Core, and ToR (Top of Rack). The 10G SFP+ AOC is used for the interconnection between the ToR access switches and the server NICs. The 40G QSFP+ SR4 optical transceivers and MTP/MPO cables are used for the interconnection between the ToR access switches and the Edge Core switches. The 100G QSFP28 CWDM4 optical transceivers and duplex LC cables are used for the interconnection between the Edge Core switches and the Spine Core switches.

 

 

The trend of the port bandwidth upgrades for Cloud Data Centers is from 10G to 25G, and then from 25G to 100G.

 

Upgrade Path 2008–2014 2013–2019 2017–2021 2019~
Data Center Campus 40G-LR4

40G-LR4

100G-CWDM4

100G-CWDM4 400G-FR4
Intra-Building

40G-eSR4

4x10G-SR

40G-eSR4 100G-SR4 400G-DR4
Intra-Rack CAT6 10G AOC 25G AOC 100G AOC
Sever Data Rate 1G 10G 25G 100G


According to the difference in the rate of increase in flow, network architecture, reliability requirements and the environment of the machine room compared to telecommunication networks, the demand for optical transceivers of cloud data centers has the following characteristics: shorter iteration period, higher speed, higher density, lower power consumption and use by mass.

 

Shorter iteration period. The rapid growth of data center traffic is driving the upgrading of optical transceivers in acceleration. The iteration period of data center hardware devices including optical transceivers is about 3 years, while the iterative period of telecommunication optical transceivers is usually over 6-7 years.


Higher speed. Because of the explosive growth of data center traffic, the technology iteration of optical transceivers can not catch up with the demand, and almost all the most advanced technologies are applied to data centers. For higher speed optical transceivers, there is always a demand for data centers, and the key question is whether the technology is mature or not.
Higher density. The high density core is to improve the transmission capacity of the switches and the single boards of the servers, in essence, to meet the demand of high speed increasing flow. At the same time, the higher the density is, the less switches are needed be deployed, and the resources of the machine room can be saved.


Lower power consumption. The power consumption of the data center is very large. Lower power consumption is to save energy and ensure better heat dissipation. Because there are full of optical transceivers on the backboards of the data centers, if the heat dissipation problem can not be properly solved, the performance and density of the optical transceivers will be affected.
Gigalight's cloud data center solution includes optical transceivers and active optical cables for 10G/25G/40G/100G/200G/400G networks.

 

Optical Transceivers Solution Applications Maximum Connection Distance
400G QSFP-DD SR8 400GE, 2x200GE 100m
200G QSFP-DD SR8 2x100GE 100m
200G QSFP-DD PSM8 2x100GE 2km-10km
200G QSFP56 SR4 200GE 100m
100G QSFP28 SR4 100GE, OTU4, 128GFC/4x32GFC 100m-300m
100G QSFP28 PSM4 100GE 2km-10km
100G QSFP28 CWDM4 100GE 2km-10km
100G QSFP28 CLR4 100GE 2km-10km
100G QSFP28 4WDM-10 100GE 10km
100G QSFP28 LR4 100GE, OTU4 10km-20km
100G QSFP28 4WDM-40 100GE 30km-40km
100G QSFP28 ER4 Lite 100GE, OTU4 30km-40km
50G SFP56 SR 50GE 100m
40G QSFP+ SR4 40GE, OTU3 400m
40G QSFP+ PSM4 40GE, OTU3 2km-10km
40G QSFP+ LR4 40GE, OTU3 2km-10km
40G QSFP+ ER4 40GE, OTU3 40km
     
     

 

Active Optical Cables Solution Applications Maximum Connection Distance
400G QSFP-DD AOC 400GE, 2x200GE 100m
200G QSFP-DD/QSFP56 AOC 2x100GE, 200GE 100m
100G QSFP28 AOC 100GE, 128GFC/4x32GFC 100m-300m
50G SFP56 AOC 50GE 100m
40G/56G QSFP AOC 40GE, 4x16GFC, 2x25GE, 2x32GFC 150m-300m
25G SFP28 AOC 25GE, 32GFC 100m-300m
10G SFP+ AOC 10GE 300m