华为数据中心3.0架构白皮书
- 格式:pdf
- 大小:1.16 MB
- 文档页数:8
VEA华为:虚拟现实(VR)体验标准技术白皮书虚拟现实(VR)体验标准技术白皮书
白皮书定义了VR体验评估的三个维度:视听沉浸体验质量、观看体验质量(业务连续性)和交互体验质量,基于ITU规范进行了样本获取和数据建模,并最终形成客观模型。
该白皮书的发布是VR体验评估模型标准化进程中的重要里程碑。
本白皮书核心洞察:VR体验,终端创新是基础,内容应用是核心,网络平
台是保障。
以终端创新为基础,进一步减轻重量提升佩戴舒适性,进一步提高屏幕分辨率和扩大FOV;以内容应用为核心,提升VR视频内容从8K全景起步,提高内容帧率到60fps及以上来缓解眩晕,3D立体景深视觉效果是体验刚需;以
运营商网络和Cloud VR平台为保障,在光纤和5G双千兆网络高带宽、低时延能力支持下,提供云渲染能力,加快VR产业大规模应用进程。
具体在VEA视频体验联盟秘书处及工作组总体规划,在核心贡献单位华为X Labs实验室、西安电子科技大学多媒体实验室,组织单位华为GTS、首都师范大学机器视觉实验室负责具体组织和实施,并在广大成员单位积极参与主观测试下,领域专家和工程师系统研究VR体验关键影响因素,构建了评估模型框架,并通过符合ITU规范的主观实验方法,阶段性地建立起VR体验评估模型。
评估模型从视听沉浸体验质量、观看体验质量和交互体验质量的角度,对分辨率、帧率、码率、FOV、MTP、自由度等20多个指标进行定义、建模和量化,为满足运营商、设备制造商、内容提供商等在不同层面、不同角度的体验评估需求,提供不同的评估模块,本白皮书是VR体验评估标准化进程的重要里程碑。
当然,现阶段受限于主观测试数据规模和指标测量手段,后续还需要VEA视频体验联盟和产业伙伴继续通力合作,提升模型准确性,并对模型实施更新迭代。
CloudEngine 8800&7800&6800 系列交换机FCoE 和DCB 技术白皮书目录1FC SAN 简介 (1)2FCoE 简介 (2)3FC 原理描述 (5)3.1FC 的基本概念 (6)3.2FC 的工作原理 (8)3.3FCF (8)3.4NPV (9)3.5Zone (10)4FCoE 原理描述 (12)4.1FCoE 的基本概念 (13)4.2FCoE 的封装 (15)4.3FIP 协议 (16)4.4FSB (19)5应用场景 (22)5.1FCF 场景 (23)5.2FCF+NPV 组网 (24)5.3FCF+FSB 组网 (26)6配置任务概览 (28)7配置注意事项 (29)8兼容性列表 (33)9缺省配置 (53)10配置FCF. (54)10.1配置FC/FCoE 接口 (57)10.2(可选)配置FC 接口参数 (58)10.3配置FCF 实例 (59)10.4配置Zone (59)10.5(可选)配置FCF 参数 (61)10.6(可选)指定FCoE 流量端口 (61)10.7(可选)配置FIP Keepalive 报文的发送间隔 (63)10.8检查配置结果 (63)11配置NPV. (64)11.1配置FC/FCoE 接口 (67)11.2(可选)配置FC 接口参数 (68)11.3配置NPV 实例 (69)11.4(可选)配置接口映射关系 (70)11.5(可选)配置FIP Keepalive 报文的发送间隔 (70)11.6检查配置结果 (71)12配置FSB (72)12.1配置FSB 实例 (74)12.2配置接口角色 (74)12.3(可选)配置FIP 协议报文的交互超时时间 (75)12.4(可选)配置FCoE 链路同步功能 (76)12.5(可选)使能协议报文隔离功能 (76)12.6检查配置结果 (77)13维护FCoE (78)13.1查看FIP 报文的统计信息 (79)13.2清除FIP 报文的统计信息 (79)13.3查看FC 和FIP 报文的统计信息 (79)13.4清除FC 和FIP 报文的统计信息 (80)13.5监控FCoE 的运行状态 (80)14配置举例 (81)14.1配置FCF 示例(FC 接口) (82)14.2配置FCF 示例(FCoE 接口) (85)14.3配置FCF 示例(FC 和FCoE 接口) (89)14.4配置FCF+NPV 示例(FCoE 接口) (93)14.5配置FCF+NPV 示例(FC 和FCoE 接口) (100)14.6配置FCF+FSB 示例(FCoE 接口) (105)15DCB 简介 (112)16DCB 原理描述 (113)16.1 PFC. (114)16.2 ETS (116)16.3 DCBX (118)17 应用 (122)18配置任务概览 (123)19配置注意事项 (124)20缺省配置 (126)21配置PFC 功能 (127)22配置ETS 功能 (129)22.1配置ETS 模板 (130)22.2应用ETS 模板 (131)22.3检查配置结果 (131)23配置DCBX 功能 (132)24维护DCB (134)24.1监控DCB 的运行状态 (135)24.2清除DCB 的统计信息 (135)25配置举例 (136)25.1配置DCB 示例 (137)1FC SAN 简介介绍FC SAN 的定义和作用。
01041.1 电信网络发展的机遇与挑战1.2 电信产业自动驾驶网络探索与实践08 1.3 华为自动驾驶网络探索实践1114 2.2 华为自动驾驶网络目标架构011. 电信自动驾驶网络探索与实践112. 华为自动驾驶网络战略与架构2.1 华为自动驾驶网络战略CONTENTS目录694. 自动驾驶网络产业发展建议745. 总结24263.1 华为自动驾驶网络解决方案3.2 网络极简系列产品35 3.3 iMaster智能运维系列产品243. 华为自动驾驶网络解决方案和产品75 6. 参考文献Page 01电信自动驾驶网络探索与实践电信自动驾驶网络探索与实践在历史发展的滚滚长河中,人类追求先进生产力的脚步从不曾停歇。
每一次技术革命的出现,都代表了一次生产力的发展更迭,驱动人类社会迈向新的发展纪元。
工业革命,电力革命和信息技术革命在过去120年间实现了人类文明的三次巨大突破,当今,以人工智能、5G、云计算为主导的第四次工业革命所带来的改变,已在悄然发生,正在塑造一个万物感知、万物互联、万物智能的世界,它比我们想象中更快地到来。
电信网络发展的机遇与挑战华为展望2025年行业发展趋势,大带宽、低时延、广链接的需求正在驱动5G 加速商用,到2025年”全天候实时在线”将成为人与万物的“默认状态”。
GIV预测2025年全球将部署650万基站服务于28亿用户58%的人口将享有5G服务650万28亿58%Page 02电信自动驾驶网络探索与实践蓬勃发展的万物互联和融合智能化给电信产业的ICT投资带来了新的发展机遇,同时也带来了挑战。
根据OVUM分析报告显示,过去十年电信行业的收入增长从来没有跑赢过OPEX的增长,随着网络规模逐年增加,OPEX快速增长,产业结构化矛盾日益突出。
无线工厂将持续演进和发展,智能自动化在建筑、制作、医疗健康等领域中广泛应用。
与此同时,智能云和云边协同技术将成为全社会运行的基础环境。
随着行业数据的不断丰富,AI算法和学习也将不断升级。
华为使能水务行业数字化智慧水务之旅智慧水务持续建设与发展01数字经济正在强势崛起,已成为未来经济发展的主要动力,如何在未来10年最大化收获数字化溢出带来的经济增长硕果,成为各行各业都非常关心的话题。
如何利用新一代ICT 信息技术,实现以创新为核心的数字化转型是水务未来发展的重要途径。
水务行业目前迎来数字化转型的快速发展期,国家不断出台新政策,促进水务行业信息化技术应用整体水平稳步提高,从2019年水利部鄂竟平部长提出水利发展改革总基调“补短板、强监管”,到2020年水利网信工作要点提出实施网络安全能力提升,完善水利信息化基础设施,积极推进5G 技术在水利工程安全监测预警等工作中的应用,信息化已经成为国家水利部重点工作之一,智慧水务建设逐步成为时代发展新模式。
智慧水务发展背景1.1 趋势:技术驱动,行业转型,势在必行近年来水利信息化建设虽然取得了较大成绩,但和补短板,强监管的要求相比还存在各种差距。
首先是底层透彻感知不够,感知覆盖范围和要素内容不全面,在绝大部分中小河流监测和水库等场景均缺乏相关监控措施,业务工作基本是巡护靠走,识别靠瞅。
现存水利信息化系统多为分散构建模式,众多信息孤岛,内外部信息共享不足,对互联网数据利用深度不够。
业务应用覆盖面不全,应用智能化水平不够,部分业务流程未能固化,还处于线下办理,体外循环。
同时,水务系统整体支撑演进能力不足,无法发挥高新技术的潜能,急需利用大数据、人工智能、遥感等技术支持水务业务的创新,让更多业务应用能够基于人工智能识别等技术实现分析型应用。
水务建设还面临信息系统安全防护及保障问题。
据统计,全国省级以上水利部门的众多应用系统中仅不足30%通过等级保护测评,大型水利工程控制系统核心设备和软件大多存在安全隐患,尚未实现以国产化为主。
相关装备、网络安全、应用支撑、系统运维等技术和管理标准还有待完善,改善目前重建设、轻运维的现状,实现对信息化建设成果的继承性,提升应用效果。
华为TaiShan 200 服务器产品白皮书(型号 2180)目录1产品概述 (1)2产品特点 (2)3逻辑结构 (4)4硬件描述 (6)4.1外观 (6)4.2指示灯和按钮 (9)4.3Riser 卡和PCIe 槽位 (14)4.4物理结构 (17)5产品规格 (19)5.1技术规格 (19)5.2环境规格 (21)5.3物理规格 (22)6软硬件兼容性 (24)6.1CPU (24)6.2 内存 (25)6.3 存储 (28)6.4 IO 扩展 (29)6.5 电源 (29)7系统管理 (30)8维保与保修 (32)9通过的认证 (33)1 产品概述TaiShan 200 服务器是基于华为鲲鹏920 处理器的数据中心服务器,其中,2180 是2U1路均衡型机架服务器(以下简称2180)。
该服务器面向互联网、分布式存储、云计算、大数据、企业业务等领域,具有高性能计算、大容量存储、低能耗、易管理、易部署等优点。
外观图如图1-1 所示。
图1-1 外观图2 产品特点性能和扩展特点2180 的性能和扩展特点如下:●支持华为自研的、面向服务器领域的64 bits 高性能多核华为鲲鹏920 处理器,内部集成了DDR4、PCIe4.0、25GE、10GE、GE 等接口,提供完整的SOC 功能。
−最大可支持64cores,2.6GHz,可支持多种核数量和频率的型号搭配。
−兼容适配ARMv8-A 架构特性,支持ARMv8.1 和ARMv8.2 扩展。
−Core 为自研64bits-TaiShan core 核。
−每个core 集成64KB L1 ICache,64KB L1 Dcache 和512KB L2 Dcache。
−支持高达45.5~46MB 的L3 cache 容量。
−支持超标量,可变长度,乱序流水线。
−支持ECC 1bit 纠错,ECC 2bit 报错。
−支持8 个DDR 控制器。
−最大支持4 个物理以太网口。
FusionStorage 技术白皮书FusionStorage 8.0 技术白皮书目录目录1概述 (1)2产品价值 (2)2.1分布式存储,随需而用 (2)2.2以弹性高效满足关键业务数据存储需求 (3)2.3丰富的企业级特性,助您构建高可用数据中心 (3)2.4开放兼容,下一代云基础设施的理想选择 (3)2.5智能数据服务与系统运维管理 (3)3产品架构 (5)3.1相关概念 (5)3.2软件架构 (6)3.3硬件架构 (7)3.4网络架构 (8)3.4.1以太网组网方案 (8)3.4.1.1部署方式 (8)3.4.1.2计算和存储分离部署 (8)3.4.1.3计算和存储融合部署 (11)3.4.2InfiniBand 组网方案 (13)3.4.2.1部署方式 (13)3.4.2.2计算和存储分离部署 (13)3.4.2.3计算和存储融合部署 (15)3.4.3RoCE 组网方案 (17)3.4.3.1部署方式 (17)3.4.3.2计算和存储分离部署 (17)3.4.3.3计算和存储融合部署 (19)3.5关键服务流程 (21)3.5.1组件描述 (21)3.5.2访问协议 (22)3.5.3数据路由 (23)3.5.4读IO 流程 (24)3.5.5写IO 流程 (24)4块存储特性 (26)4.1精简配置 (26)4.2重删压缩 (26)4.3多资源池 (28)4.4数据加密 (29)4.5QoS (30)4.6 快照 (32)4.7 克隆 (34)4.8异步复制 (34)4.9AA 双活 (35)5弹性扩展 (37)5.1DHT 算法 (37)5.2平滑扩容 (39)5.3性能扩展 (40)6高性能 (42)6.1分布式存储优化算法 (42)6.1.1动态智能分区和静态选盘算法 (42)6.1.2快速编码快速重构的弹性EC 算法 (43)6.2分布式SSD Cache 加速 (45)6.2.1Write Cache (46)6.2.2Read Cache (47)6.2.3大IO Pass Through (48)6.2.4动态Cache 调整 (49)7安全性 (50)7.1安全框架 (50)7.2设备安全 (51)7.3网络安全 (51)7.4业务安全 (52)7.5管理安全 (52)8可靠性 (53)8.1硬件可靠性 (53)8.2软件可靠性 (54)8.2.1节点冗余设计 (54)8.2.2网络链路聚合 (54)8.2.3亚健康管理 (55)8.2.3.1硬盘亚健康管理 (55)8.2.3.2网络亚健康管理 (56)8.2.3.3服务亚健康管理 (56)8.2.3.4快速换路重试 (56)8.3数据可靠性 (56)8.3.1数据保护 (56)8.3.1.1多副本 (57)8.3.1.2纠删码 (57)8.3.1.3多故障域 (58)8.3.1.4掉电保护 (59)8.3.1.5快速数据重建 (59)8.3.2数据一致性 (60)8.3.2.1强一致性复制协议 (60)8.3.2.2读修复技术 (60)8.3.2.3数据完整性保护 (60)8.4解决方案可靠性 (61)8.4.1本地数据保护 (61)8.4.2业务连续性保护 (62)8.4.2.1AA 双活 (62)8.4.2.2异步复制 (62)9开放兼容性 (64)9.1存储协议兼容性 (64)9.2虚拟化平台兼容性 (64)9.3云管平台兼容性 (65)9.3.1OpenStack 云管平台 (65)9.3.2非OpenStack 云管平台 (65)9.4综合网管平台兼容性 (65)9.5软件兼容性 (66)9.5.1操作系统兼容性 (66)9.5.2数据库软件兼容性 (66)9.5.3大数据应用兼容性 (66)9.6硬件兼容性 (66)9.6.1服务器硬件 (66)9.6.2存储介质兼容性 (67)9.6.3IO 板卡兼容性 (67)10存储永新 (68)10.1存储服务更新 (68)10.2系统滚动升级 (69)10.3新硬件替换 (69)10.3.1老平台纳管 (69)10.3.2新硬件加入 (70)11存储管理 (71)11.1块存储服务 (71)11.2块存储集群管理 (72)11.3eSight 数据中心级管理 (72)11.4eService 云化管理 (73)11.5SmartKit 智能巡检 (73)12缩略语 (74)1 概述随着数据不断增长以及互联网业务的兴起,新兴应用对存储需求的快速变化以及不确定性成为主要挑战,在金融行业,银行要抓住互联网、特别是移动互联网金融崛起带来的机遇,同时也不得不迎接由此带来的挑战:新业务天级甚至小时级上线;更精准的用户需求分析等。
Technical White PaperHigh Throughput Computing Data Center ArchitectureThinking of Data Center 3.0 AbstractIn the last few decades, data center (DC) technologies have kept evolving fromDC 1.0 (tightly-coupled silos) to DC 2.0 (computer virtualization) to enhance dataprocessing capability. Emerging big data analysis based business raiseshighly-diversified and time-varied demand for DCs. Due to the limitations onthroughput, resource utilization, manageability and energy efficiency, current DC2.0 shows its incompetence to provide higher throughput and seamlessintegration of heterogeneous resources for different big data applications. Byrethinking the demand for big data applications, Huawei proposes a highthroughput computing data center architecture (HTC-DC). Based on resourcedisaggregation and interface-unified interconnects, HTC-DC is enabled withPB-level data processing capability, intelligent manageability, high scalability andhigh energy efficiency. With competitive features, HTC-DC can be a promisingcandidate for DC3.0.ContentsEra of Big Data: New Data Center Architecture in Need 1⏹Needs on Big Data Processing 1⏹DC Evolution: Limitations and Strategies 1⏹Huawei’s Vision on Future DC 2DC3.0: Huawei HTC-DC 3⏹HTC-DC Overview 3⏹Key Features 4Summary 6June 2014ERA OF BIG DATA: NEW DATA CENTER ARCHITECTURE IN NEED⏹Needs on Big Data Processing During the past few years, applications which arebased on big data analysis have emerged, enrichinghuman life with more real-time and intelligentinteractions. Such applications have proven themselvesto become the next wave of mainstream of onlineservices. As the era of big data approaches, higher andhigher demand on data processing capability has beenraised. Being the major facilities to support highlyvaried big data processing tasks, future data centers(DCs) are expected to meet the following big datarequirements (Figure 1):▪PB/s-level data processing capability ensuring aggregated high-throughput computing, storage and networking; ▪Adaptability to highly-varied run-time resource demands; ▪Continuous availability providing 24x7 large-scaled service coverage, and supporting high-concurrency access; ▪ Rapid deployment allowing quick deployment and resource configuration for emerging applications.⏹ DC Evolution: Limitations and StrategiesDC technologies in the last decade have been evolved (Figure 2) from DC 1.0 (with tightly-coupled silos) to current DC 2.0 (with computer virtualization). Although data processing capability of DCs have been significantly enhanced, due to the limitations on throughput, resource utilization, manageability and energy efficiency, current DC 2.0 shows its incompetence to meet the demands of the future:Figure 2. DC Evolution- Throughput: Compared with technological improvement in computational capability of processors, improvement in I/O access performance has long been lagged behind. With the fact that computing within conventional DC architecture largely involves data movement between storage and CPU/memory via I/O ports, it is challenging for current DC architecture to provide PB-level high throughput for big data applications. The problem of I/O gap is resulted from low-speed characteristics of conventional transmission and storage mediums, and also from inefficient architecture design and data access mechanisms.To meet the requirement of future high throughput data processing capability, adopting new transmission technology (e.g. optical interconnects) and new storage medium can be feasible solutions. But a more fundamental approach is to re-design DC architecture as well as data access mechanisms for computing. If data access in computing process can avoid using conventional I/O mechanism, but use ultra-high-bandwidth network to serve as the new I/O functionality, DC throughput can be significantly improved.Figure 1. Needs Brought by Big DataJune 2014- Resource Utilization:Conventional DCs typically consist of individual servers which are specifically designed for individual applications with various pre-determined combinations of processors, memories and peripherals. Such design makes DC infrastructure very hard to adapt to emergence of various new applications, so computer virtualization technologies are introduced accordingly. Although virtualization in current DCs help improve hardware utilization, it cannot make use of the over-fractionalized resource, and thus making the improvement limited and typically under 30%1,2. As a cost, high overhead exists with hypervisor which is used as an essential element when implementing computer virtualization. In addition, in current DC architecture, logical pooling of resources is still restricted by the physical coupling of in-rack hardware devices. Thus, current DC with limited resource utilization cannot support big data applications in an effective and economical manner.One of the keystones to cope with such low utilization problem is to introduce resource disaggregation, i.e., decoupling processor, memory, and I/O from its original arrangements and organizing resources into shared pools. Based on disaggregation, on-demand resource allocation and flexible run-time application deployment can be realized with optimized resource utilization, reducing Total Cost of Operation (TCO) of infrastructure.- Manageability: Conventional DCs only provide limited dynamic management for application deployment, configuration and run-time resource allocation. When scaling is needed in large-scaled DCs, lots of complex operations still need to be completed manually.To avoid complex manual re-structuring and re-configuration, intelligent self-management with higher level of automation is needed in future DC. Furthermore, to speed up the application deployment, software defined approaches to monitor and allocate resources with higher flexibility and adaptability is needed.- Energy Efficiency: Nowadays DCs collectively consume about 1.3% of all Array global power supply3. As workload of big data drastically grows, future DCswill become extremely power-hungry. Energy has become a top-lineoperational expense, making energy efficiency become a critical issue in greenDC design. However, the current DC architecture fails to achieve high energyefficiency, with the fact that a large portion of energy is consumed for coolingother than for IT devices.With deep insight into the composition of DC power consumption (Figure3), design of each part in a DC can be more energy-efficient. To identify andeliminate inefficiencies and then radically cut energy costs, energy-savingdesign of DC should be top-to-bottom, not only at the system level but also atFigure 3. DC Power Consumption the level of individual components, servers and applications.Huawei’s Vision on Future DCIn Huawei’s vision, to support future big data applications, future DCs should be enabled with the following features:- Big-Data-Oriented: Different from conventional computing-centric DCs, data-centric should be the key design concept of DC3.0. Big data analysis based applications have highly varied characteristics, based on which DC 3.0 should provide optimizedmechanisms for rapid transmission, highly concurrent processing of massive data, and also for application-diversified acceleration.- Adaptation for Task Variation: Big data analysis brings a booming of new applications, raising different resource demands that vary with time. In addition, applications have different need for resource usage priority. To meet such demand variation with highadaptability and efficiency, disaggregation of hardware devices to eliminate the in-rack coupling can be a key stone. Such a methodenables flexible run-time configuration on resource allocation, ensuring the satisfactory of varied resource demand of different applications.- Intelligent Management: DC 3.0 involves massive hardware resource and high density run-time computation, requiring higher intelligent management with less need for manual operations. Application deployment and resource partitioning/allocation, even system diagnosis need to be conducted in automated approaches based on run-time monitoring and self-learning. Further, Service Level Agreement (SLA) guaranteeing in complex DC computing also requires a low-overhead run-time self-manageable solution.1./index.cfm?c=power_mgt.datacenter_efficiency_consolidation2./system-optimization/a-data-center-conundrum/3./green/bigpicture/#/datacenters/infographicsJune 2014- High Scalability: Big data applications require high throughput low-latency data access within DCs. At the same time, extremely high concentration of data will be brought into DC facilities, driving DCs to grow into super-large-scaled with sufficient processing capability. It is essential to enable DCs to maintain acceptable performance level when ultra-large-scaling is conducted.Therefore, high scalability should be a critical feature that makes a DC design competitive for the big data era.- Open, Standard based and Flexible Service Layer: With the fact that there exists no unified enterprise design for dynamical resource management at different architecture or protocol layers, from IO, storage to UI. Resources cannot be dynamically allocated based on the time and location sensitive characteristics of the application or tenant workloads. Based on the common principles of abstraction and layering, open and standard based service-oriented architecture (SOA) has been proven effective and efficient and has enabled enterprises of all sizes to design and develop enterprise applications that can be easily integrated and orchestrated to match their ever-growing business and continuous process improvement needs, while software defined networking (SDN) has also been proven in helping industry giants such as Google to improve its DC network resource utilization with decoupling of control and data forwarding, and centralized resource optimization and scheduling. To provide competitive big data related service, an open, standard based service layer should be enabled in future DC to perform application driven optimization and dynamic scheduling of the pooled resources across various platforms.- Green: For future large-scale DC application in a green and environment friendly approach, energy efficient components, architectures and intelligent power management should be included in DC 3.0. The use of new mediums for computing, memory, storage and interconnects with intelligent on-demand power supply based on resource disaggregation help achieving fine-grained energy saving. In addition, essential intelligent energy management strategies should be included: 1) Tracking the operational energy costs associated with individual application-related transactions; 2) Figuring out key factors leading to energy costs and conduct energy-saving scheduling; 3) Tuning energy allocation according to actual demands; 4) Allowing DCs to dynamically adjust the power state of servers, and etc.DC3.0: HUAWEI HTC-DCHTC-DC OverviewTo meet the demands of high throughput in the big data era, current DC architecture suffers from critical bottlenecks, one of which is the difficulty to bridge the I/O performance gap between processor and memory/peripherals. To overcome such problem and enable DCs with full big-data processing capability, Huawei proposes a new high throughput computing DC architecture (HTC-DC), which avoids using conventional I/O mechanism, but uses ultra-high-bandwidth network to serve as the new I/O functionality. HTC-DC integrates newly-designed infrastructures based on resource disaggregation, interface-unified interconnects and a top-to-bottom optimized software stack. Big data oriented computing is supported by series of top-to-bottom accelerated data operations, light weighted management actions and the separation of data and management.Figure 4. Huawei HTC-DC ArchitectureFigure 4 shows the architecture overview of HTC-DC. Hardware resources are organized into different pools, which are links up together via interconnects. Management plane provides DC-level monitoring and coordination via DC Operating System (OS), while business-related data access operations are mainly conducted in data plane. In the management plane, a centralized ResourceJune 2014Figure 5. Hardware Architecture of Huawei HTC-DC Management Center (RMC) conducts global resource partitioning/allocation and coordination/scheduling of the related tasks, with intelligent management functionalities such as load balancing, SLA guaranteeing, etc. Light-hypervisor provides abstract of pooled resources, and performs lightweight management that focuses on execution of hardware partitioning and resource allocation but not get involved in data access. Different from conventional hypervisor which includes data access functions in virtualization, light-hypervisor focuses on resource management, reducing complexity and overhead significantly. As a systematical DC3.0 design, HTC-DC also provides a complete software stack to support various DC applications. A programming framework with abundant APIs is designed to enable intelligent run-time self-management.Key FeaturesResource Disaggregated Hardware SystemFigure 5 illustrates the hardwarearchitecture of HTC-DC, which isbased on completely-disaggregatedresource pooling. The computing poolis designed with heterogeneity. Eachcomputing node (i.e. a board) carriesmultiple processors (e.g., x86, Atom,Power and ARM, etc.) for application-diversified data processing. Nodes inmemory pool adopt hybrid memorysuch as DRAM and non-volatilememory (NVM) for optimized high-throughput access. In I/O pool,general-purposed extension (GPU,massive storage, external networking,etc.) can be supported via differenttypes of ports on each I/O node. Eachnode in the three pools is equippedwith a cloud controller which canconduct diversified on-board manage-ment for different types of nodes. Pooled Resource Access Protocol (PRAP)To form a complete DC, all nodes in the three pools are interconnected via a network based on a new designed Pooled Resource Access Protocol (PRAP). To reduce the complexity of DC computing, HTC-DC introduces PRAP which has low-overhead packet format, RDMA-enabled simplified protocol stack, unifying the different interfaces among processor, memory and I/O. PRAP is implemented in the cloud controller of each node to provide interface-unified interconnects. PRAP supports hybrid flow/packet switching for inter-pool transmission acceleration, with near-to-ns latency. QoS can be guaranteed via run-time bandwidth allocation and priority-based scheduling. With simplified sequencing and data restoring mechanisms, light-weight lossless node-to-node transmission can be achieved.With resource disaggregation and unified interconnects, on-demand resource allocation can be supported by hardware with fine-granularity, and intelligent management can be conducted to achieve high resource utilization (Figure 6). RMC in the management plane provides per-minute based monitoring, on-demand coordination and allocation over hardware resources. Required resources from the pools can be appropriately allocated according to the characteristics of applications (e.g. Hadoop). Optimized algorithm assigns and schedules tasks on specific resource partitions where customized OSs are hosted. Thus, accessibility and bandwidth of remote memory and peripherals can be ensured within the partition, and hence end-to-end SLA can be guaranteed. Enabled with self-learning mechanisms, resource allocation and management in HTC-DC requires minimal manual operation, bringing intelligence and efficiency.June 2014Figure 6. On-demand Resource Allocation Based on DisaggregationHuawei Many-Core Data Processing UnitTo increase computing density, uplift data throughput and reduce communication latency, Huawei initializes Data Processing Unit (DPU, Figure 7) which adopts lightweight-core based many-core architecture, heterogeneous 3D stacking and Through-Silicon Vias (TSV) technologies. In HTC-DC, DPU can be used as the main computing component. The basic element of DPU is Processor-On-Die (POD), which consists of NoC, embedded NVM, clusters with heavy/light cores, and computing accelerators. With software-defined technologies, DPU supports resource partitioning and QoS-guaranteed local/remote resource sharing that allow application to directly access resources within its assigned partition. With decoupled multi-threading support, DPU executes speculative tasks off the critical path, resulting in enhanced overall performance. Therefore static power consumptions can be significantly reduced. Especially, some of the silicon chip area can be saved by using the optimal combinations of the number of synchronization and execution pipelines, while maintaining the same performance.Figure 7. Many-Core ProcessorNVM Based StorageEmerging NVM (including MRAM or STT-RAM, RRAM and PCM, etc.) has been demonstrated with superior performance over flash memories. Compared to conventional storage mediums (hard-disk, SSD, etc.), NVM provides more flattened data hierarchy with simplified layers, being essential to provide sufficient I/O bandwidth. In HTC-DC, NVMs are employed both as memory and storage. NVM is a promising candidate for DRAM replacement with competitive performance but lower power consumption. When used as storage, NVM provides 10 times higher IOPS than SSD4, bringing higher data processing capability with enhanced I/O performance.Being less hindered by leakage problems with technology scaling and meanwhile having a lower cost of area, NVM is being explored extensively to be the complementary medium for the conventional SDRAM memory, even in L1 caches. Appropriately tuning of selective architecture parameters can reduce the performance penalty introduced by the NVM to extremely tolerable levels while obtaining over 30% of energy gains.54./global/business/semiconductor/news-events/press-releases/detail?newsId=129615.M. Komalan et.al., “Feasibility exploration of NVM based I-cache through MSHR enhancements”, Proceeding in DATE’14June 2014Optical InterconnectsTo meet the demand brought by big data applications, DCs are driven to increase the data rate on links (>10Gbps) while enlarging the scale of interconnects (>1m) to host high-density components with low latency. However due to non-linear power consumption and signal attenuation, conventional copper based DC interconnects cannot have competitive performance with optical interconnects on signal integrity, power consumption, form factor and cost6. In particular, optical interconnect has the advantage of offering large bandwidth density with low attenuation and crosstalk. Therefore a re-design of DC architecture is needed to fully utilize advantages of optical interconnects. HTC-DC enables high-throughput low-latency transmission with the support of interface-unified optical interconnects. The interconnection network of HTC-DC employs low-cost Tb/s-level throughput optical transceiver and co-packaged ASIC module, with tens of pJ/bit energy consumption and low bit error rate for hundred-meter transmission. In addition, with using intra/inter-chip optical interconnects and balanced space-time-wavelength design, physical layer scalability and the overall power consumption can be enhanced. Using optical transmission that needs no signal synchronization, PRAP-based interconnects provide higher degree of freedom on topology choosing, and is enabled to host ultra-large-scale nodes.DC-Level Efficient Programming FrameworkTo fully exploit the architectural advantages and provide flexible interface for service layer to facilitate better utilization of underlying hardware resource, HTC-DC provides a new programming framework at DC-level. Such a framework includes abundant APIs, bringing new programming methodologies. Via these APIs, applications can issue requests for hardware resource based on their demands. Through this, optimized OS interactions and self-learning-based run-time resource allocation/scheduling are enabled.In addition, the framework supports automatically moving computing operations to near-data nodes while keeping data transmission locality. DC overhead is minimized by introducing topology-aware resource scheduler and limiting massive data movement within the memory pool.In addition, Huawei has developed the Domain Specific Language (HDSL) as part of the framework to reduce the complexity of programming in HTC-DC for parallelism. HDSL includes a set of optimized data structures with operations (such as Parray, parallel processing the data in array) and a parallel processing library. One of the typical applications of HDSL is for graph computing. HDSL can enable efficient programming with competitive performance. Automated generation of distributed codes is also supported.SUMMARYWith the increasing growth of data consumption, the age of big data brings new opportunities as well as great challenges for future DCs. DC technologies have been evolved from DC 1.0 (tightly-coupled server) to DC 2.0 (software virtualization) with the data processing capability being largely enhanced. However, the limited I/O throughput, energy inefficiency, low resource utilization and limited scalability of DC 2.0 become the bottlenecks to fulfill big data application demand. Therefore, a new, green and intelligent DC 3.0 architecture fitting different resource demands of various big-data applications is in need.With the design avoiding data access via conventional I/O but using ultra-high-bandwidth network to serve as the new I/O functionality, Huawei proposes HTC-DC as a new generation of DC design for future big data applications. HTC-DC architecture enables a DC to compute as a high throughput computer. Based on the resource disaggregated architecture and interface-unified PRAP network, HTC-DC integrates many-core processor, NVM, optical interconnects and DC-level efficient programming framework.Such a DC ensures PB-level data processing capability, supporting intelligent management, being easy and efficient to scale, and significantly saves energy.HTC-DC architecture is still being developed. Using Huawei’s cutting-edge technologies, HTC-DC can be a promising candidate design for the future, ensuring a firm step for DCs to head for the big data era.■6.“Silicon Photonics Market & Technologies 2011-2017: Big Investments, Small Business”, Yole Development, 2012Copyright © Huawei Technologies Co., Ltd. 2014. All rights reserved.No part of this document may be reproduced or transmitted in any form or by any means without prior written consent of Huawei Technologies Co., Ltd.and other Huawei trademarks are trademarks of Huawei Technologies Co., Ltd.All other trademarks and trade names mentioned in this document are the property of their respective holders.NoticeThe information in this document may contain predictive statements including, without limitation, statements regarding the future financial and operating results, future product portfolio, new technology, etc. There are a number of factors that could cause actual results and developments to differ materially from those expressed or implied in the predictive statements. Therefore, such information is provided for reference purpose only and constitutes neither an offer nor an acceptance. Huawei may change the information at any time without notice.Huawei Technologies Co., Ltd.Address: Huawei Industrial BaseBantian, LonggangShenzhen 518129People's Republic of ChinaWebsite: Tel: 0086-755-28780808。