v200f - T-Platforms
v200f - T-Platforms
v200f - T-Platforms
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
V-CLASS FAMILY UPDATE<br />
V5050 | V200S | V200SS | V200F2 | V200FS | V402S<br />
JUNE 2014<br />
Public presentation<br />
VER 1 | ENG
AGENDA<br />
• RECENT V-CLASS CHANGES<br />
• T-PLATFORMS SYSTEMS’ PORTOFLIO<br />
• V-CLASS POSITIONING<br />
• V5050(L) ENCLOSURE<br />
• INTEL XEON E5-2600 v2-BASED MODULES<br />
• V402S SERVER MODULE (VLIW)<br />
• V200 APPLICATION MATRIX<br />
• MODULE PORTFOLIO MATRIX<br />
• MANAGEMENT SUBSYSTEM<br />
• SUMMARY<br />
• BACKUP INFORMATION<br />
2
RECENT V-CLASS CHANGES<br />
• New V5050(L) enclosure<br />
• V5000 and V5050L are at stock<br />
• V5050 are built to order<br />
• New server modules<br />
• V200SS – Intel Xeon E5-2600 v2-based server with hot- and cold-swap<br />
disks<br />
• V200FS – Intel Xeon E5-2600 v2 ---based server with fixed, hot- and<br />
cold-swap disks<br />
• V402S – MCST Elbrus-based server with cold-swap disks<br />
• AMD Opteron 6300-based modules are phased out<br />
• V205S – built to large orders only<br />
• V205F – discontinued<br />
3
SYSTEMS’ PORTFOLIO<br />
3 rd party board<br />
T-MINI P (PRODUCTION):<br />
• Minicluster/WS systems<br />
• 4-8 DP compute trays<br />
• 1 fixed head<br />
• INTEL//NVIDIA<br />
• Integrated storage<br />
• Integrated networking<br />
Workgroup<br />
Shared T-<strong>Platforms</strong> V210 system board and<br />
common management<br />
E-CLASS (H2’14):<br />
• 2U general<br />
server/storage server<br />
• 1 DP compute node<br />
• V210 board (INTEL)<br />
• HS disks<br />
• Shares V-Class<br />
management<br />
V-CLASS (PRODUCTION):<br />
• 5U air-cooled chassis<br />
• 10 DP compute modules<br />
• 5 accelerated DP<br />
modules<br />
• V2x0/V205/402 boards<br />
INTEL/AMD/NVIDIA/ELBRUS<br />
• Centralized management<br />
Workgroup, Departmental and Divisional<br />
A-CLASS (2014):<br />
• ~52U cabinet<br />
• Hot water cooling<br />
• Integrated PSU blades<br />
• Integrated blade<br />
switches<br />
• 1P INTEL+ NVIDIA GPU<br />
@ launch<br />
Rack level integration<br />
Supercomputer<br />
FAMILY ABBREVIATION SEGMENT<br />
A-CLASS ADVANCED HIGH-END HPC<br />
V-CLASS VOLUME MIDDLE-END HPC/HIGH END CLOUD/SMB<br />
E-CLASS ENTERPRISE STORAGE OR GENERAL SERVER<br />
T-MINI P (P-CLASS) PERSONAL SCALABLE ALL-IN ONE HPC/WS SYSTEMS<br />
4
V-CLASS POSITIONING<br />
• Scalable server family for<br />
datacenters and internet<br />
service providers:<br />
• Heterogeneous HPC<br />
• Remote visualization,<br />
rendering desktop<br />
virtualization<br />
• Infrastructure and hosting<br />
applications of service<br />
providers<br />
• Application servers<br />
5
V5050(L) ENCLOSURE<br />
• 5U chassis for 19″ racks with min depth of 1070mm<br />
• Rear bays for 5-10 modules<br />
• ‘Mix and match’ various servers<br />
• Hot-plug support for cooling modules. PSUs and servers<br />
• Integrated management module with Ethernet switch and iKVM<br />
• LED control panel<br />
• RS-485 interface to connect to the AESS (V5050)<br />
• 2 GbE-порта управления шасси (V5050)<br />
• RS-232 serial interface<br />
• Up to 5 CRPS '80Plus Platinum‘ power supplies<br />
• N+2, N+1 redundancy level support<br />
• Up to 6.4kW of supplied power with N+1 redundancy<br />
• 3 Highly Efficient N+1 cooling modules (2 coaxial fans each)<br />
6
V5050(L) ENCLOSURE (2)<br />
Cooling module<br />
Enclosure front<br />
• No cabled connections<br />
• Mid-plane is a passive design with<br />
mating connector<br />
• Enclosure with PSUs – 38.35kg<br />
Indication and control panel<br />
Enclosure with no servers, rear view<br />
• V200S:<br />
• V200SS:<br />
• V200F2:<br />
• V200FS:<br />
~ 4.9 kg (2 disks)<br />
~ 6.8 kg (4 disks)<br />
~ 8.7 kg (2 disks)<br />
~ 13.1 kg (10 disks)<br />
• The enclosure is supplied with fixed<br />
rack-mount rails<br />
CRPS Platinum<br />
PSU<br />
Management ports and<br />
UID (V5050 version)<br />
7
V5050L & V5050 DIFFERENCES<br />
V5000 V5050L V5050<br />
PSUs 4 x PSUs 5 x PSUs (CRPS)<br />
Raiting Platinum Platinum (до 94%)<br />
V5050L with 5 х V200FS<br />
PSU<br />
redundancy<br />
Management<br />
ports<br />
N+1 N+1/N+2<br />
1 x GbE 2 x GbE<br />
AESS ports -- -- RS485<br />
UID and error<br />
code<br />
reporting<br />
-- -- Present<br />
V5050 with 5 х V200FS<br />
8
2-WAY INTEL XEON MODULES<br />
• 4 passively cooled V200 modules<br />
• V200S – compute module with cold-swap disks<br />
• V200SS – server module with cold- and hot-swap disks<br />
• V200F2 – double-width HPC/visualization module with 2 GPUs and coldswap<br />
disks<br />
• V200FS – double-width storage-enhanced server module<br />
• Common feature set<br />
• Up to two Intel Xeon E5-2600 v2 CPUs (TDP - 135W, except for V200SS<br />
with 95W TDP)<br />
• 16 slots for REG DDR3 DIMM, up to 1866 MHz<br />
• 2 integrated GbE ports, optional 10/40GbE/FDR InfiniBand port<br />
• 2 cold-swap 2.5” HDDa or SSDs (cable-free connection)<br />
• Internal and external management interfaces<br />
• ASPEED 2150 BMC<br />
9
V200S MODULE<br />
• Standard compute module,<br />
supporting two 2.5” cold-swap disks<br />
• Chipset-integrated SAS/SATA with<br />
RAID 0/1<br />
• 2 disks use 3Gbit/s SAS interface<br />
• The best choice for high<br />
performance compute clusters and<br />
server farms with centralized storage<br />
• Supports one low profile PCIe x16 Gen.3 adapter<br />
• Fanless tray has no cabled connections inside<br />
• A configuration of 10 х V200S reaches 4.3 Tflops of peak<br />
performance<br />
10
V200F2 MODULE<br />
• Double-width compute module,<br />
supporting two passively cooled<br />
accelerators and two 2.5” cold-swap<br />
disks<br />
• Chipset-integrated SAS/SATA with RAID 0/1<br />
• 2 disks use 3Gbit/s SAS interface<br />
• Accelerators are supported through two PCIe<br />
x16 Gen.3 slots equipped with riser cards<br />
• NVIDIA Tesla K10, K20, K20X, K40<br />
• NVIDIA GRID K1 & K2<br />
• Intel Xeon® Phi 3120P, 5110P & 7120P<br />
• Best suited for accelerated parallel computing, remote visualization and client<br />
virtualization<br />
• The HPC configuration of 5 х V200F2 reaches over 18 Tflops of peak<br />
performance<br />
11
THE NEW V200SS MODULE<br />
• Standard server module, supporting<br />
two 2.5” hot-swap disks and two 2.5”<br />
cold-swap disks<br />
• Chipset-integrated SAS/SATA with<br />
RAID 0/1<br />
• Hot-swap disks are using SATA 6 Gbit/s<br />
• Cold-swap disks are using SAS 3 Gbit/s<br />
• Positioned for broader range of server applications<br />
• This module has a CPU thermal design power limit at 95W<br />
• Fanless tray has cabled connections inside<br />
12
THE NEW V200FS MODULE<br />
• Double-width server module<br />
supporting 4 hot-swap and 4<br />
cold-swap 2.5” disks and 4<br />
fixed 2.5” or 3.5” disks<br />
• Chipset-integrated SAS/SATA<br />
with RAID 0/1<br />
• Supports a single 4- or 8-port<br />
RAID adapter or HBA though<br />
PCIe x16 Gen.3 slot<br />
• Check out the V200FS online section for various storage<br />
subsystem configuration options<br />
• Total local storage space can reach 36TB<br />
• Fanless tray has cabled connections inside<br />
13
THE NEW V200FS MODULE (2)<br />
14
THE NEW V402S MODULE<br />
• Standard width server module featuring four<br />
heterogeneous MCST Elbrus-2C+ CPUs (25W<br />
TDP)<br />
• Designed for customers interested in VLIW<br />
architecture potential and for embedded<br />
applications<br />
• One module carries 8 core VLIW cores and 32<br />
DSPs<br />
• One PCIe x8 Gen.1.0a expansion slot<br />
• Fanless cable-free tray<br />
• Comes in 10 X V402S or mixed configurations<br />
• Built to order<br />
V402S<br />
15
V200-SERIES APPLICATION MATRIX<br />
APPLICATION TYPE<br />
RATIN<br />
G<br />
HPC CLUSTER ***<br />
RECOMMENDED<br />
MODULES<br />
V200S & V200F2 w<br />
NVIDIA Tesla K/Intel<br />
Xeon Phi<br />
NOTES<br />
WEB-SERVER (FRONT END) ** V200S/V200SS More expensive than 1U-server<br />
WEB-SERVER (MIDDLE END) *** V200SS /V200FS<br />
WEB-SERVER<br />
(BACK END) ---<br />
Low automation, insufficient availability, only 4 hot swap disks<br />
out of 10 , low number of PCIe slots, unnecessary density<br />
CLOUD APPS ** V200S/V200SS More expensive than 1U-servers and custom platforms<br />
REMOTE VISUALIZATION ***<br />
V200F2 w NVIDIA GRID<br />
K1/K2<br />
CItrix-ready (XenServer 6.2), NVIDIA GRID K-certified<br />
TERMINAL SERVER *** V200S CItrix-ready (XenServer 6.2)<br />
APPLICATION SERVER *** V200SS and V200FS<br />
SMB-SERVER --- The system is too dense and costly<br />
DATABASE SERVER * V200FS<br />
VIRTUALIZATION ** V200S<br />
Low automation, insufficient availability, only 4 hot swap disks<br />
out of 10 , low number of PCIe slots, unnecessary density<br />
No virtual server support by system’s management, actual<br />
XenServer 6.2 certification, expecting VMWare certification<br />
16
V-CLASS MODULE MATRIX<br />
Model V200S V200F2 V200SS V200FS V402S V205S V205F<br />
Module type Standard Double Standard Double Standard Standard Double<br />
CPU 2 x Intel Xeon 2 x E5-2600 v2 4 x Elbrus 2С+ 2 х AMD Opteron 6300<br />
Memory<br />
16 х Reg DIMM DDR3, up to 1866MHz<br />
8 х Reg<br />
miniDIMM<br />
DDR2<br />
16 х Reg DIMM DDR3, up to<br />
1600МГц<br />
СS disks 2 2 2 2 2 2 2<br />
HS disks - - 2 4 - - -<br />
FX disks - - -<br />
PCIe slots<br />
1 x PCIe x16<br />
Gen 3<br />
2 x PCIe x16<br />
Gen 3<br />
-<br />
4 (2.5” or<br />
3.5”)<br />
1 x PCIe x16<br />
Gen 3<br />
- - -<br />
1 x PCIe x8<br />
Gen 1<br />
1 x PCIe x16 Gen 2<br />
LP adapters 1 - 1 1 1 1 -<br />
Accelerators<br />
(FH FL)<br />
- 2 - - - - 1 (M2090)<br />
GbE ports Integrated 2 x GbE (Intel I350) Int., 1 x GbE Integrated, 2 x GbE<br />
Integrated IB<br />
port<br />
10/40GbE or FDR InfiniBand ( 1 QSFP with optional SFP+<br />
adapter)<br />
10GbE/1 x QDR InfiniBand<br />
(QSFP)<br />
BMC<br />
Integrated AST2150<br />
miniDIMM with<br />
AST2400<br />
Integrated AST2050<br />
CS – cold-swap<br />
HS – hot-swap<br />
FX – fixed<br />
LP – low profile<br />
FH FL – full size, full lenght<br />
17
CONFIGURATION NOTES<br />
Up to 10 V200S, V200SS or V402S<br />
Up to 5 V200F2 or V200FS<br />
Mixed configurations<br />
Blank modules<br />
• All spare bays have to host blank modules<br />
to maintain correct airflow<br />
18
SYSTEM’S MANAGEMENT (1)<br />
SYSTEM’S MANAGEMENT CONTROLLER<br />
Node control panel x 10<br />
Fans and PSU status LEDs<br />
Chassis power button<br />
Chassis Front Section<br />
• Indication and control module - 1U, cold-swappable<br />
Serial/GbE ports to control all nodes<br />
• Chassis and nodes management ports consolidation (one or two cables for 5-10 nodes)<br />
• One independent GbE port on chassis front and two at the back (one is reserved for future use)<br />
• Node indication, power and UID buttons | PSU and cooling modules status<br />
Inside the Systems’ Management Module<br />
• ARM-based minicomputer (SoC)<br />
• Management switch with 11 internal 100Mbit/s ports<br />
and three external GbE ports for node management<br />
consolidation<br />
• iKVM<br />
19
SYSTEM’S MANAGEMENT(2)<br />
V5050(L) features integrated web server with GUI(IMU) for in-band<br />
and out-of-band system deployment and updating:<br />
• Configuration detection<br />
• Management network setup for SMC и modules’ BMC<br />
• Pass-through access to modules’ BMC<br />
• Monitoring (temp, tachs, power)<br />
• Reset control<br />
• FRU<br />
• SEL<br />
• Remote flashing<br />
• Backup and firmware rollback<br />
• Various access levels<br />
20
SYSTEM’S MANAGEMENT (3)<br />
Web-интерфейс IMU 1.2<br />
21
SYSTEM’S MANAGEMENT(4)<br />
Compute and server modules carry ASPEED AST2150 BMC (a<br />
miniDIMM with AST2400 for the V402S) with integrated VGA и<br />
two MACs<br />
• Full IPMI 2.0 support, including KVM over LAN, Remote<br />
Media, remote monitoring and management<br />
• Secure WEB and CLI-based access<br />
• Connection to compute module individual BMCs via systems<br />
management controller or directly via GbE port 1<br />
• Remote firmware update for BIOS and BMC<br />
• No need for special utilities or pre-installed OS<br />
miniDIMM with<br />
BMC for V402S<br />
22
SUMMARY<br />
• Great variety of available compute and server modules (Intel, AMD,<br />
MCST)<br />
• High performance of Intel Xeon-based modules<br />
• Up to 4.3 Tflops per system with 10 x V200S (TDP 135 W)<br />
• Over 18 Tflops per system with 5 x V200F2 carrying 10 NVIDIA Tesla K40 GPUs<br />
• Support for DDR3-1866MHz configurations<br />
• Energy efficiency<br />
• Optimized airflow<br />
• Platinum power supplies<br />
• Optimized cooling fans rotation algorithms<br />
• Easy to deploy and manage<br />
• Consolidated system management<br />
• Intel ‘Haswell’ E5-2600 v3 modules are planned for H2’2014<br />
23
THANK YOU!<br />
t-platforms.com/v-class<br />
sales@t-platforms.ru<br />
24
BACKUP SLIDES<br />
25
INTERNAL LAYOUT (V5000)<br />
Compute modules with 2.5”<br />
disks<br />
Cooling modules<br />
PSUs<br />
Power distribution<br />
board<br />
Midplane<br />
Control panel<br />
Management controller
BLOCK CHART OF V5050L<br />
27
BLOCK CHART OF V5050<br />
28
V200 SYSTEM BOARD<br />
4<br />
5<br />
7<br />
3<br />
2<br />
8<br />
9<br />
7<br />
6<br />
6<br />
1 1<br />
6<br />
6<br />
10<br />
11<br />
12<br />
Узлы V200S/F2<br />
используют<br />
унифицированную<br />
системную плату<br />
Системная плата V200-B2A<br />
CPU’S<br />
EXPANSION SLOTS AND PORTS<br />
1. Socket 2011 for Intel Xeon E5-2600 v2 6. 16 DIMM DDR 3 slots<br />
7. 2 PCIe x16 Gen.3 slots<br />
INTEGRATED CONTROLLERS<br />
8. 2 miniSAS: 4 SAS, 4 SATA<br />
9 1 USB2.0 Type A port<br />
2. Intel PCH C600 chipset<br />
3. 1-port Mellanox ConnectX-3 chip (an option)<br />
4. 2-port Intel I350 GbE controller<br />
5. ASPEED 2150 BMC/VGA<br />
SPECIALIZED INTERFACES<br />
10. CardEdge (for identity card to interface with midplane)<br />
11. CardEdge SAS 1 for one CS disk<br />
12. CardEdge SAS 2 for one CS disk<br />
29
ADDITIONAL IMAGES<br />
Chassis rear<br />
Management controller extraction Cooling modules Power supplies (V5000 version)<br />
Chassis front<br />
30
ADDITIONAL IMAGES (2)<br />
V200F with 2 NVIDIA® Tesla GPUs<br />
V200S<br />
* Modules are shown without a top panel<br />
31
ADDITIONAL IMAGES (3)<br />
• V-Class-based HPC cluster<br />
• Cold-door cabinet<br />
• 7 x V5000 with V200S and<br />
V200F2 modules<br />
32
V-CLASS USE CASES<br />
Stony Brook University, USA, 2012<br />
CSC-IT for success, Finland, 2012<br />
Juelich Research Centre, Germany, 2013<br />
Over 25 projects in Russia, 2012-14<br />
Подготовка стойки с<br />
холодной дверью<br />
для центра Юлих,<br />
Германия<br />
33