10.05.2015 Views

v200f - T-Platforms

v200f - T-Platforms

v200f - T-Platforms

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

V-CLASS FAMILY UPDATE<br />

V5050 | V200S | V200SS | V200F2 | V200FS | V402S<br />

JUNE 2014<br />

Public presentation<br />

VER 1 | ENG


AGENDA<br />

• RECENT V-CLASS CHANGES<br />

• T-PLATFORMS SYSTEMS’ PORTOFLIO<br />

• V-CLASS POSITIONING<br />

• V5050(L) ENCLOSURE<br />

• INTEL XEON E5-2600 v2-BASED MODULES<br />

• V402S SERVER MODULE (VLIW)<br />

• V200 APPLICATION MATRIX<br />

• MODULE PORTFOLIO MATRIX<br />

• MANAGEMENT SUBSYSTEM<br />

• SUMMARY<br />

• BACKUP INFORMATION<br />

2


RECENT V-CLASS CHANGES<br />

• New V5050(L) enclosure<br />

• V5000 and V5050L are at stock<br />

• V5050 are built to order<br />

• New server modules<br />

• V200SS – Intel Xeon E5-2600 v2-based server with hot- and cold-swap<br />

disks<br />

• V200FS – Intel Xeon E5-2600 v2 ---based server with fixed, hot- and<br />

cold-swap disks<br />

• V402S – MCST Elbrus-based server with cold-swap disks<br />

• AMD Opteron 6300-based modules are phased out<br />

• V205S – built to large orders only<br />

• V205F – discontinued<br />

3


SYSTEMS’ PORTFOLIO<br />

3 rd party board<br />

T-MINI P (PRODUCTION):<br />

• Minicluster/WS systems<br />

• 4-8 DP compute trays<br />

• 1 fixed head<br />

• INTEL//NVIDIA<br />

• Integrated storage<br />

• Integrated networking<br />

Workgroup<br />

Shared T-<strong>Platforms</strong> V210 system board and<br />

common management<br />

E-CLASS (H2’14):<br />

• 2U general<br />

server/storage server<br />

• 1 DP compute node<br />

• V210 board (INTEL)<br />

• HS disks<br />

• Shares V-Class<br />

management<br />

V-CLASS (PRODUCTION):<br />

• 5U air-cooled chassis<br />

• 10 DP compute modules<br />

• 5 accelerated DP<br />

modules<br />

• V2x0/V205/402 boards<br />

INTEL/AMD/NVIDIA/ELBRUS<br />

• Centralized management<br />

Workgroup, Departmental and Divisional<br />

A-CLASS (2014):<br />

• ~52U cabinet<br />

• Hot water cooling<br />

• Integrated PSU blades<br />

• Integrated blade<br />

switches<br />

• 1P INTEL+ NVIDIA GPU<br />

@ launch<br />

Rack level integration<br />

Supercomputer<br />

FAMILY ABBREVIATION SEGMENT<br />

A-CLASS ADVANCED HIGH-END HPC<br />

V-CLASS VOLUME MIDDLE-END HPC/HIGH END CLOUD/SMB<br />

E-CLASS ENTERPRISE STORAGE OR GENERAL SERVER<br />

T-MINI P (P-CLASS) PERSONAL SCALABLE ALL-IN ONE HPC/WS SYSTEMS<br />

4


V-CLASS POSITIONING<br />

• Scalable server family for<br />

datacenters and internet<br />

service providers:<br />

• Heterogeneous HPC<br />

• Remote visualization,<br />

rendering desktop<br />

virtualization<br />

• Infrastructure and hosting<br />

applications of service<br />

providers<br />

• Application servers<br />

5


V5050(L) ENCLOSURE<br />

• 5U chassis for 19″ racks with min depth of 1070mm<br />

• Rear bays for 5-10 modules<br />

• ‘Mix and match’ various servers<br />

• Hot-plug support for cooling modules. PSUs and servers<br />

• Integrated management module with Ethernet switch and iKVM<br />

• LED control panel<br />

• RS-485 interface to connect to the AESS (V5050)<br />

• 2 GbE-порта управления шасси (V5050)<br />

• RS-232 serial interface<br />

• Up to 5 CRPS '80Plus Platinum‘ power supplies<br />

• N+2, N+1 redundancy level support<br />

• Up to 6.4kW of supplied power with N+1 redundancy<br />

• 3 Highly Efficient N+1 cooling modules (2 coaxial fans each)<br />

6


V5050(L) ENCLOSURE (2)<br />

Cooling module<br />

Enclosure front<br />

• No cabled connections<br />

• Mid-plane is a passive design with<br />

mating connector<br />

• Enclosure with PSUs – 38.35kg<br />

Indication and control panel<br />

Enclosure with no servers, rear view<br />

• V200S:<br />

• V200SS:<br />

• V200F2:<br />

• V200FS:<br />

~ 4.9 kg (2 disks)<br />

~ 6.8 kg (4 disks)<br />

~ 8.7 kg (2 disks)<br />

~ 13.1 kg (10 disks)<br />

• The enclosure is supplied with fixed<br />

rack-mount rails<br />

CRPS Platinum<br />

PSU<br />

Management ports and<br />

UID (V5050 version)<br />

7


V5050L & V5050 DIFFERENCES<br />

V5000 V5050L V5050<br />

PSUs 4 x PSUs 5 x PSUs (CRPS)<br />

Raiting Platinum Platinum (до 94%)<br />

V5050L with 5 х V200FS<br />

PSU<br />

redundancy<br />

Management<br />

ports<br />

N+1 N+1/N+2<br />

1 x GbE 2 x GbE<br />

AESS ports -- -- RS485<br />

UID and error<br />

code<br />

reporting<br />

-- -- Present<br />

V5050 with 5 х V200FS<br />

8


2-WAY INTEL XEON MODULES<br />

• 4 passively cooled V200 modules<br />

• V200S – compute module with cold-swap disks<br />

• V200SS – server module with cold- and hot-swap disks<br />

• V200F2 – double-width HPC/visualization module with 2 GPUs and coldswap<br />

disks<br />

• V200FS – double-width storage-enhanced server module<br />

• Common feature set<br />

• Up to two Intel Xeon E5-2600 v2 CPUs (TDP - 135W, except for V200SS<br />

with 95W TDP)<br />

• 16 slots for REG DDR3 DIMM, up to 1866 MHz<br />

• 2 integrated GbE ports, optional 10/40GbE/FDR InfiniBand port<br />

• 2 cold-swap 2.5” HDDa or SSDs (cable-free connection)<br />

• Internal and external management interfaces<br />

• ASPEED 2150 BMC<br />

9


V200S MODULE<br />

• Standard compute module,<br />

supporting two 2.5” cold-swap disks<br />

• Chipset-integrated SAS/SATA with<br />

RAID 0/1<br />

• 2 disks use 3Gbit/s SAS interface<br />

• The best choice for high<br />

performance compute clusters and<br />

server farms with centralized storage<br />

• Supports one low profile PCIe x16 Gen.3 adapter<br />

• Fanless tray has no cabled connections inside<br />

• A configuration of 10 х V200S reaches 4.3 Tflops of peak<br />

performance<br />

10


V200F2 MODULE<br />

• Double-width compute module,<br />

supporting two passively cooled<br />

accelerators and two 2.5” cold-swap<br />

disks<br />

• Chipset-integrated SAS/SATA with RAID 0/1<br />

• 2 disks use 3Gbit/s SAS interface<br />

• Accelerators are supported through two PCIe<br />

x16 Gen.3 slots equipped with riser cards<br />

• NVIDIA Tesla K10, K20, K20X, K40<br />

• NVIDIA GRID K1 & K2<br />

• Intel Xeon® Phi 3120P, 5110P & 7120P<br />

• Best suited for accelerated parallel computing, remote visualization and client<br />

virtualization<br />

• The HPC configuration of 5 х V200F2 reaches over 18 Tflops of peak<br />

performance<br />

11


THE NEW V200SS MODULE<br />

• Standard server module, supporting<br />

two 2.5” hot-swap disks and two 2.5”<br />

cold-swap disks<br />

• Chipset-integrated SAS/SATA with<br />

RAID 0/1<br />

• Hot-swap disks are using SATA 6 Gbit/s<br />

• Cold-swap disks are using SAS 3 Gbit/s<br />

• Positioned for broader range of server applications<br />

• This module has a CPU thermal design power limit at 95W<br />

• Fanless tray has cabled connections inside<br />

12


THE NEW V200FS MODULE<br />

• Double-width server module<br />

supporting 4 hot-swap and 4<br />

cold-swap 2.5” disks and 4<br />

fixed 2.5” or 3.5” disks<br />

• Chipset-integrated SAS/SATA<br />

with RAID 0/1<br />

• Supports a single 4- or 8-port<br />

RAID adapter or HBA though<br />

PCIe x16 Gen.3 slot<br />

• Check out the V200FS online section for various storage<br />

subsystem configuration options<br />

• Total local storage space can reach 36TB<br />

• Fanless tray has cabled connections inside<br />

13


THE NEW V200FS MODULE (2)<br />

14


THE NEW V402S MODULE<br />

• Standard width server module featuring four<br />

heterogeneous MCST Elbrus-2C+ CPUs (25W<br />

TDP)<br />

• Designed for customers interested in VLIW<br />

architecture potential and for embedded<br />

applications<br />

• One module carries 8 core VLIW cores and 32<br />

DSPs<br />

• One PCIe x8 Gen.1.0a expansion slot<br />

• Fanless cable-free tray<br />

• Comes in 10 X V402S or mixed configurations<br />

• Built to order<br />

V402S<br />

15


V200-SERIES APPLICATION MATRIX<br />

APPLICATION TYPE<br />

RATIN<br />

G<br />

HPC CLUSTER ***<br />

RECOMMENDED<br />

MODULES<br />

V200S & V200F2 w<br />

NVIDIA Tesla K/Intel<br />

Xeon Phi<br />

NOTES<br />

WEB-SERVER (FRONT END) ** V200S/V200SS More expensive than 1U-server<br />

WEB-SERVER (MIDDLE END) *** V200SS /V200FS<br />

WEB-SERVER<br />

(BACK END) ---<br />

Low automation, insufficient availability, only 4 hot swap disks<br />

out of 10 , low number of PCIe slots, unnecessary density<br />

CLOUD APPS ** V200S/V200SS More expensive than 1U-servers and custom platforms<br />

REMOTE VISUALIZATION ***<br />

V200F2 w NVIDIA GRID<br />

K1/K2<br />

CItrix-ready (XenServer 6.2), NVIDIA GRID K-certified<br />

TERMINAL SERVER *** V200S CItrix-ready (XenServer 6.2)<br />

APPLICATION SERVER *** V200SS and V200FS<br />

SMB-SERVER --- The system is too dense and costly<br />

DATABASE SERVER * V200FS<br />

VIRTUALIZATION ** V200S<br />

Low automation, insufficient availability, only 4 hot swap disks<br />

out of 10 , low number of PCIe slots, unnecessary density<br />

No virtual server support by system’s management, actual<br />

XenServer 6.2 certification, expecting VMWare certification<br />

16


V-CLASS MODULE MATRIX<br />

Model V200S V200F2 V200SS V200FS V402S V205S V205F<br />

Module type Standard Double Standard Double Standard Standard Double<br />

CPU 2 x Intel Xeon 2 x E5-2600 v2 4 x Elbrus 2С+ 2 х AMD Opteron 6300<br />

Memory<br />

16 х Reg DIMM DDR3, up to 1866MHz<br />

8 х Reg<br />

miniDIMM<br />

DDR2<br />

16 х Reg DIMM DDR3, up to<br />

1600МГц<br />

СS disks 2 2 2 2 2 2 2<br />

HS disks - - 2 4 - - -<br />

FX disks - - -<br />

PCIe slots<br />

1 x PCIe x16<br />

Gen 3<br />

2 x PCIe x16<br />

Gen 3<br />

-<br />

4 (2.5” or<br />

3.5”)<br />

1 x PCIe x16<br />

Gen 3<br />

- - -<br />

1 x PCIe x8<br />

Gen 1<br />

1 x PCIe x16 Gen 2<br />

LP adapters 1 - 1 1 1 1 -<br />

Accelerators<br />

(FH FL)<br />

- 2 - - - - 1 (M2090)<br />

GbE ports Integrated 2 x GbE (Intel I350) Int., 1 x GbE Integrated, 2 x GbE<br />

Integrated IB<br />

port<br />

10/40GbE or FDR InfiniBand ( 1 QSFP with optional SFP+<br />

adapter)<br />

10GbE/1 x QDR InfiniBand<br />

(QSFP)<br />

BMC<br />

Integrated AST2150<br />

miniDIMM with<br />

AST2400<br />

Integrated AST2050<br />

CS – cold-swap<br />

HS – hot-swap<br />

FX – fixed<br />

LP – low profile<br />

FH FL – full size, full lenght<br />

17


CONFIGURATION NOTES<br />

Up to 10 V200S, V200SS or V402S<br />

Up to 5 V200F2 or V200FS<br />

Mixed configurations<br />

Blank modules<br />

• All spare bays have to host blank modules<br />

to maintain correct airflow<br />

18


SYSTEM’S MANAGEMENT (1)<br />

SYSTEM’S MANAGEMENT CONTROLLER<br />

Node control panel x 10<br />

Fans and PSU status LEDs<br />

Chassis power button<br />

Chassis Front Section<br />

• Indication and control module - 1U, cold-swappable<br />

Serial/GbE ports to control all nodes<br />

• Chassis and nodes management ports consolidation (one or two cables for 5-10 nodes)<br />

• One independent GbE port on chassis front and two at the back (one is reserved for future use)<br />

• Node indication, power and UID buttons | PSU and cooling modules status<br />

Inside the Systems’ Management Module<br />

• ARM-based minicomputer (SoC)<br />

• Management switch with 11 internal 100Mbit/s ports<br />

and three external GbE ports for node management<br />

consolidation<br />

• iKVM<br />

19


SYSTEM’S MANAGEMENT(2)<br />

V5050(L) features integrated web server with GUI(IMU) for in-band<br />

and out-of-band system deployment and updating:<br />

• Configuration detection<br />

• Management network setup for SMC и modules’ BMC<br />

• Pass-through access to modules’ BMC<br />

• Monitoring (temp, tachs, power)<br />

• Reset control<br />

• FRU<br />

• SEL<br />

• Remote flashing<br />

• Backup and firmware rollback<br />

• Various access levels<br />

20


SYSTEM’S MANAGEMENT (3)<br />

Web-интерфейс IMU 1.2<br />

21


SYSTEM’S MANAGEMENT(4)<br />

Compute and server modules carry ASPEED AST2150 BMC (a<br />

miniDIMM with AST2400 for the V402S) with integrated VGA и<br />

two MACs<br />

• Full IPMI 2.0 support, including KVM over LAN, Remote<br />

Media, remote monitoring and management<br />

• Secure WEB and CLI-based access<br />

• Connection to compute module individual BMCs via systems<br />

management controller or directly via GbE port 1<br />

• Remote firmware update for BIOS and BMC<br />

• No need for special utilities or pre-installed OS<br />

miniDIMM with<br />

BMC for V402S<br />

22


SUMMARY<br />

• Great variety of available compute and server modules (Intel, AMD,<br />

MCST)<br />

• High performance of Intel Xeon-based modules<br />

• Up to 4.3 Tflops per system with 10 x V200S (TDP 135 W)<br />

• Over 18 Tflops per system with 5 x V200F2 carrying 10 NVIDIA Tesla K40 GPUs<br />

• Support for DDR3-1866MHz configurations<br />

• Energy efficiency<br />

• Optimized airflow<br />

• Platinum power supplies<br />

• Optimized cooling fans rotation algorithms<br />

• Easy to deploy and manage<br />

• Consolidated system management<br />

• Intel ‘Haswell’ E5-2600 v3 modules are planned for H2’2014<br />

23


THANK YOU!<br />

t-platforms.com/v-class<br />

sales@t-platforms.ru<br />

24


BACKUP SLIDES<br />

25


INTERNAL LAYOUT (V5000)<br />

Compute modules with 2.5”<br />

disks<br />

Cooling modules<br />

PSUs<br />

Power distribution<br />

board<br />

Midplane<br />

Control panel<br />

Management controller


BLOCK CHART OF V5050L<br />

27


BLOCK CHART OF V5050<br />

28


V200 SYSTEM BOARD<br />

4<br />

5<br />

7<br />

3<br />

2<br />

8<br />

9<br />

7<br />

6<br />

6<br />

1 1<br />

6<br />

6<br />

10<br />

11<br />

12<br />

Узлы V200S/F2<br />

используют<br />

унифицированную<br />

системную плату<br />

Системная плата V200-B2A<br />

CPU’S<br />

EXPANSION SLOTS AND PORTS<br />

1. Socket 2011 for Intel Xeon E5-2600 v2 6. 16 DIMM DDR 3 slots<br />

7. 2 PCIe x16 Gen.3 slots<br />

INTEGRATED CONTROLLERS<br />

8. 2 miniSAS: 4 SAS, 4 SATA<br />

9 1 USB2.0 Type A port<br />

2. Intel PCH C600 chipset<br />

3. 1-port Mellanox ConnectX-3 chip (an option)<br />

4. 2-port Intel I350 GbE controller<br />

5. ASPEED 2150 BMC/VGA<br />

SPECIALIZED INTERFACES<br />

10. CardEdge (for identity card to interface with midplane)<br />

11. CardEdge SAS 1 for one CS disk<br />

12. CardEdge SAS 2 for one CS disk<br />

29


ADDITIONAL IMAGES<br />

Chassis rear<br />

Management controller extraction Cooling modules Power supplies (V5000 version)<br />

Chassis front<br />

30


ADDITIONAL IMAGES (2)<br />

V200F with 2 NVIDIA® Tesla GPUs<br />

V200S<br />

* Modules are shown without a top panel<br />

31


ADDITIONAL IMAGES (3)<br />

• V-Class-based HPC cluster<br />

• Cold-door cabinet<br />

• 7 x V5000 with V200S and<br />

V200F2 modules<br />

32


V-CLASS USE CASES<br />

Stony Brook University, USA, 2012<br />

CSC-IT for success, Finland, 2012<br />

Juelich Research Centre, Germany, 2013<br />

Over 25 projects in Russia, 2012-14<br />

Подготовка стойки с<br />

холодной дверью<br />

для центра Юлих,<br />

Германия<br />

33

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!