01.02.2013 Views

Chapter 4 PA-RISC Computer Systems - OpenPA.net

Chapter 4 PA-RISC Computer Systems - OpenPA.net

Chapter 4 PA-RISC Computer Systems - OpenPA.net

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Convex Exemplar SPP1000, SPP1200 & SPP1600 Internals<br />

� SPP1000/XA Cluster: 8-128 <strong>PA</strong>-7100 100 MHz with 1/1 MB off-chip I/D L1 cache each<br />

� SPP1200/CD: 2-16 <strong>PA</strong>-7200 120 MHz with 256/256 KB off-chip I/D L1 cache each<br />

� SPP1200/XA Hypernode: 2-8 <strong>PA</strong>-7200 120 MHz with 256/256 KB off-chip I/D L1 cache each<br />

� SPP1200/XA Cluster: 8-128 <strong>PA</strong>-7200 120 MHz with 256/256 KB off-chip I/D L1 cache each<br />

� SPP1600/CD: 2-16 <strong>PA</strong>-7200 120 MHz with 1/1 MB off-chip I/D L1 cache each<br />

� SPP1600/XA HyperNode: 2-16 <strong>PA</strong>-7200 120 MHz with 1/1 MB off-chip I/D L1 cache each<br />

� SPP1600/XA Cluster: 8-128 <strong>PA</strong>-7200 120 MHz with 1/1 MB off-chip I/D L1 cache each<br />

It is not quite clear how the CD models relate to the XA models — the XA clusters consist of several 2-8<br />

processor hypernodes while the CD models were shipped with up to 16 processors. Either the CDs are<br />

different machines than the XA hypernodes or they are simply two XA hypernodes coupled together,<br />

without any additional SCI/CTI expansion possibilities.<br />

Chipset<br />

The chipset is based completely on an own Convex design and centers around the Convex five-port<br />

crossbar, later improved on the SPP2000 with eight ports and used in HP’s V-Class.<br />

1. 5x5 nonblocking crossbar, with five crossbar ports, is the central part of the system, it connects<br />

to four “functional units” (memory, SCI links and processor) and with the fifth port to the local<br />

system I/O. The four functional units contain each a memory controller, SCI controller and<br />

an “agent” for two processors. Memory and processor use different data links to the crossbar<br />

— memory access always goes over the crossbar, even from a processor to the memory in the<br />

same functional unit. Each crossbar port has a data rate of 250 MB/s, giving the crossbar a combined<br />

peak bandwidth of 1.25 GB/s. The crossbar is implemented in Gallium arsenide gate arrays<br />

(GaAs, 250K transistors), quite a rarity, since it was very expensive and difficult to handle.<br />

2. Four CPU Agents attach to the crossbar and provide access for the processors to the memory via<br />

the crossbar over a 250 MB/s crossbar port shared with the memory controller (see below).<br />

3. Four Convex Coherent Memory Controllers (CCMCs) attach each one four-way interleaved<br />

memory board to the crossbar. The CCMCs additionally do cache coherency and interface to<br />

the Convex’s SCI (CTI) link for inter-hypernode connection. [It is not quite clear if the CCMCs<br />

share the whole 250 MB/s port/data connections with the CPU agents on the same functional unit,<br />

or if CCMC and CPU agent attach to separate lines of the crossbar port — Ed.] The CTI interface<br />

— or the complete CCMC — were apparently also GaA chips.<br />

4. Exemplar I/O (Input/Ouput) Subsystem connects to the fifth 250 MB/s crossbar port and attaches<br />

the I/O subsystem controllers to the crossbar and this memory and processors.<br />

Buses<br />

� Total crossbar bandwidth 1.25 GB/s (five 250 MB/s ports)<br />

� CPU/Memory bandwidth 1.0 GB/s (four 250 MB/s ports shared with memory)<br />

� I/O bandwidth 250 MB/s (one crossbar port)<br />

� SPP1000: Four SBus I/O buses for expansion slots<br />

328

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!