Professional Documents
Culture Documents
Engineered Systems
Extends
on-chip crypto
acceleration
RAS
Scales to 8 sockets
using directory
Minimizes latency
Avoids congestion
Maximize bandwidth
Dynamically thread
Optimize
Systems
Multiply
Performance
SPARC
T5
Scale
Efficiently
Advance
Power
Management
Maximizes peak
performance
Scales elastically
SPARC T5-1B
SPARC T5-4
SPARC T5-2
SPARC T5-8
SPARC T5 Servers
Product Line Overview
Processor
SPARC T5-1B
SPARC T5-2
SPARC T5-4
SPARC T5-8
SPARC T5 3.6GHz
SPARC T5
3.6GHz
SPARC T5
3.6GHz
SPARC T5
3.6GHz
16, 128
32, 256
64, 512
128, 1024
16
32
64
128
128GB or 256GB
256GB or 512GB
1TB or 2TB
2TB or 4TB
16 LP x8 PCIe 3.0,
4 x 10GbE ports
16 LP x8 PCIe 3.0,
4 x 10GbE ports
Rack 5RU
2410W
Rack 8RU
4850W
8 LP x8 PCIe 3.0,
4 x 10GbE ports
Blade
689W
Rack 3RU
1927W
Compute
1x SPARC T5 16-core CPU
16x DDR3 DIMMs
128GB (8GB DIMMs) or 256GB (16GB DIMMs)
memory
T5-1B Board
BoBs
REM
T5 CPU
PCIe
Switch
Service
Processor
16 DDR3
DIMMs
8Copyright 2012, Oracle and/or its affiliates. All rights reserved.
Compute
2x SPARC T5 16-core CPU
32x DDR3 DIMMs
256GB (8GB DIMMs) or 512GB (16GB DIMMs) of
memory
T5-2 Chassis
T5 CPU
Memory
Risers
Service
Processor
T5-2 Front
Locator
LED/Button
Power
Button
Fault LED
Status
LED
Over Temp
Indicator
Status LEDs
DISK 5
DISK 4
DISK 3
DISK 2
DISK 1
DISK 0
RFID/Serial
Number
HD-15 VGA
Port
2x USB 3.0
Ports
DVD
T5-2 Rear
HD-15 VGA
Port
SP Network 10/100
4x 10GbE
Ports
2x USB 3.0
PCIe 8
System Rear
Indicators
PCIe 7
AC0
PCIe 6
PCIe 5
PCIe 4
PCIe 3
PCIe 2
PCIe 1
PSU 0
PSU 1
AC1
SP Serial
Front
Rear
13Copyright 2012, Oracle and/or its affiliates. All rights reserved.
Feature
Form Factor
CPU
Memory
SPARC T5-4
5RU, 31.5 deep
4x SPARC T5
3.6 GHz (512 threads)
DDR3, 2TB MAX
64x Slots
Network
Internal Storage
Removable Media
Serial
PCI Express slots
Power Supply
Fans
4x 10GbE
Up to 8 x 2.5 SAS 3.0 or SSD, hot-plug
SPARC T4-4
5RU, 28 deep
4x SPARC T4
3.0 GHz (256 threads)
DDR3, 2TB MAX
64x slots
4 x 1GbE + 8x 10GbE (XAUI)
Requires 2 Separate QSFP Connectors
Up to 8 x 2.5 SAS 2.0, can use up to
4x SATA SSDs, hot-plug
Processor
Module
PM 1
PM 0
Fault LED
Status LED
Power
Button
DISK 1
DISK 0
Over Temp
Indicator
Rear Fan/EM
Indicator
RFID/Serial
Number
DISK 3
DISK 2
DISK 5
DISK 4
PSU 0
SP Serial
Port
Main
Module
DISK 7
DISK 6
PSU 1
HD-15 VGA
Port
Main Module
(Entire Board)
PSU Status
LEDs
Fan Module 2
Fan Module 3
Fan Module 4
PCIe 16
PCIe 15
PCIe 14
PCIe 13
PCIe 12
PCIe 11
PCIe 10
PCIe 9
PCIe 8
PCIe 7
PCIe 6
PCIe 5
PCIe 4
PCIe 3
PCIe 2
PCIe 1
Rear I/O
Module
(RIO)
Fan Module 1
PCIe Carrier
Hot-Plug Button, LEDs
AC3 OK
LED
AC0 OK LED
AC3
System Rear
Indicators
C19 Plug
AC0
SP Serial
4x 10GbE Ports
SP Network 10/100
2x USB 3.0
C19 Plug
Compute
8x SPARC T5 16-core CPU
128x DDR3 DIMMs
2TB (16GB DIMMs) or 4TB (32GB DIMMs) memory
Front
Rear
17Copyright 2012, Oracle and/or its affiliates. All rights reserved.
DIMMs/BoB
PCIe Gen support
SPARC T3
SPARC T4
SPARC T5
1.65GHz
2.85GHz, 3.0GHz
3.6GHz
16
16
S2
S3
2
2 BoBs/memory controller
4 DIMMs/BoB
4 BoBs/CPU socket
2.0
2 BoBs/memory controller
2 DIMMs/BoB
8 BoBs/CPU socket
3.0
T5 Processor 0
16 DIMMs per T5
16 DIMMs per T5
T5 Processor 1
CL routing
6
3
5
3
5
0 1
C1
PM0
C2
C3
C4
C5
C6
0 1
C2
C0
Part #
Sw 0
Partition #
Switch 1
Partition #
Switch 2
C6
5
C1
2
3
C3
C4
C1
C5
C6
PM3
C6 C7
0 1 0 1
PMO
C0 C1
0 1 0 1
PM1
C2 C3
0 1 0 1
C0
C7
C0
Partition #
Switch 2
Part #
Sw 4
Part #
Sw 0
Partition #
Switch 1
Partition #
Switch 2
PFM1
PFM1
C1
C2
C3
C1
PM0
C4
C5
C6
C0
C7
C0
PMO
C0 C1
0 1 0 1
PFM1
PFM2
C1
PM0
C2
C3
2
Partition #
Switch 2
1
Part #
Sw 4
0
Part #
Sw 0
3
Partition #
Switch 1
C5
C6
C7
PMO
C0 C1
0 1 0 1
PFM1
PFM2
PFM3
MP8?
3
Partition #
Switch 2
C1
MP8?
3
C4
0
4
PFM2
PFM2
MP8?
3
C7
4 5
4
2
PM0
C2
C1
PM2
C4 C5
0 1 0 1
5
2
MP8?
0
C0
C7
PM1
4
2
PMO
C0 C1
0 1 0 1
C7
4 5
PFM2
C1
2
3
CL routing
PM3
C6
5
C1
C0
5
2
C5
C0
6
2
5 4
PM1
C7
4 5
PM2
C4
C2
2
3
CL routing
PM3
C6
5
CL routing
PM3
3
Partition #
Switch 2
1
Part #
Sw 4
0
Part #
Sw 0
0
Partition #
Switch 1
0
Partition #
Switch 2
Partition #
Switch 2
Part #
Sw 4
CL routing
CL routing
PM1
6
C6
5
C0
C0
2
3
C1
C2
C3
C4
C1
PM0
PFM1
C7
4 5
C5
C6
C0
C7
C0
C1
PM0
C2
C3
Part #
Sw 0
3
Partition #
Switch 1
3
Partition #
Switch 2
C5
C6
C7
PMO
C0 C1
0 1 0 1
PFM1
MP4
3
Partition #
Switch 2
C1
MP4
0
C4
0
4
1
Part #
Sw 4
0
Part #
Sw 0
0
Partition #
Switch 1
0
Partition #
Switch 2
Partition #
Switch 2
Part #
Sw 4
BoB
BoB
BoB
CPU
DC/DCs
BoB
BoB
T5
CPU0
BoB
BoB
BoB
Host &
Data Flash
TPM
CPU
Debug
Port
PCIe
0
1
FPGA
Disk1
Disk2
x2
x1
LSI SAS
x8
x8
PCIe Gen3
PCIe Gen2
x8
x1
PCIe
Switch 0
DBG
PCIe
Switch 1
LPC
SP Module
Sideband Mgmt
x4
x8
x8
FEM0
PCI-EM1
1x4
1x4
Emulex
Pilot 3
USB 1.0
Hub Ctrl
USB
VGA
Serial
(HD15) (RJ45)
1x4
? here
22Copyright 2012, Oracle and/or its affiliates. All rights reserved.
USB 3.0
Host Ctrl
FEM1
Nalia
Niantic
Dual GigE
10/100/1000
NEM0 NEM1
x8
12C
PCI-EM0
MIDPLANE
Ethernet Mgmt
(to CMM)
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
T5-0
T5-1
CPU
Debug
Port
PCIe
CPU
DC/DCs
CPU
DC/DCs
TPM
x8
x4
x8
x1
HDD0
SAS/SATA
IO Controller
HDD0
HDD0
x1
x4
PCIe
Switch 0
HDD0
HDD0
SP Module
PCIe
Switch 1
x8
FRUID
Sideband Mgmt
x8
x8
HDD0
USB 3.0
Host
SATA DVD
USB 3.0
Hub
DBG
x4
x8
SAS/SATA
IO Controller
Host &
Data Flash
FPGA
x8
x8
CPU
Debug
Port
PCIe
DRAM
SPI
Flash
NAND
USB 2.0
Hub
USB
USB
VGA
Service
Processor
Internal USB
VGA
FAN BOARD
REAR IO
Slot 8 (8)
Slot 7 (8)
Slot 6 (8)
Slot 5 (8)
Slot 4 (8)
Slot 3 (8)
Slot 2 (8)
USB3 VGA
Slot 1 (8)
REAR IO Board
USB2
USB0 USB1
VGA
DB15
Serial
Mgmt
Enet
Mgmt
10/100
SBP1
4-DISK BACKPLANE
DISKS 4-7
MB
Motherboard
2x USB (Gen3)
[VGA/DB15]
[Serial Mgmt]
HDD[0:3]
FRONT IO
VGA, SERIAL MGMT, USB
TO RIO
SAS 2308
SAS1
4-DISK BACKPLANE
DISKS 0-3
FIO + VGA
VIDEO
MUX
Serial
MUX
SP
SERVICE PROCESSOR
MM
Main Module
SBP0
TO RIO
x2 USB Ports
SAS 2308
SAS0
Enet Mgmt
NC-SI
FRONT USB
HOST CTRLLR.
MONITOR &
CONTROL
SW1
CPU
PCIE PORTS
1 2
REAR USB
MOST CTRLLR.
10GB NIC 0
1x USB (Gen2)
2x USB (Gen3)
4 5 6
SW3
SW2
CPU
PCIE PORTS
SW5
CPU
PCIE PORTS
7 8 9 10
CPU
PCIE PORTS
11 12 13 14
SW4
Net 1
Net 0
RIO
SW0
CLOCK SYNTH.
&
BUFFERS
Debug Conn
Debug Conn
DC-DC
CONVERTERS
SW6
CPU
PCIE PORTS
EB
PCI Express Backplane
10GB NIC 1
15 16
[VGA/DB15]
[Serial Mgmt]
[Enet Mgmt 10/100]
MP4, MP8
T5-4 or T5-8
Midplane
IOS
0
IOS
1
CLR
0
CLR
1
CLR
2
CLR
3
CLR
4
CLR
5
CLR
6
IOS
0
IOS
1
CLR
0
CLR
1
T5 (CM1)
MCU0
L0
FSR0
L1
FSR1
MCU1
L0
FSR2
L1
FSR3
CLR
3
CLR
4
CLR
5
CLR
6
T5 (CM0)
MCU2
L0
FSR4
CLR
2
L1
FSR5
MCU3
L0
FSR6
L1
FSR7
MCU0
L0
FSR0
L1
FSR1
MCU1
L0
FSR2
L1
FSR3
MCU2
L0
FSR4
L1
FSR5
MCU3
L0
FSR6
L1
FSR7
BoB1
BoB3
BoB6
BoB5
BoB3
BoB0
BoB5
BoB7
C C
0 1
C C
1 0
C C
0 1
C C
1 0
C C
0 1
C C
1 0
C C
0 1
C C
1 0
BoB0
BoB2
BoB4
BoB7
BoB2
BoB1
BoB4
BoB6
C C
0 1
C C
1 0
C C
0 1
C C
1 0
C C
0 1
C C
1 0
C C
0 1
C C
1 0
T5 PCIe Subsystem
Dual x8 PCI Express Gen 3 ports provide 32 GB/s peak b/w
Supports Atomic Fetch-and-Add, Unconditional-Swap and
Compare-and-Swap operations
Accelerates virtualized I/O with Oracle Solaris VMs
128k virtual function address spaces ensure direct SR-IOV access for all
logical domains
64-bit DVMA space reduces IO mapping overhead, improving network
performance
Guarantees fault and performance isolation among guest OS instances
T5 PCIe Progression
T4
T5
16 GBs
32 GBs
44 bit
48 bit
Transaction Id Identification
on MSI and MSI-X
No
Yes
No
Yes
No
Air Flow
PCIe Retimer
x16 Connector
(x8 electrical)
28Copyright 2012, Oracle and/or its affiliates. All rights reserved.
7
Config
ID
Native Config
(T5-4 only)
PM1
CPU0
CPU1
PFM
MP
Front USB x1
Debug Slot x1
x8 up.
Part. 0
Switch 6
SAS0 x8
Switch 5
NET1 x8
x8 up.
Part. 0
x8 up.
Part. 0
x8 up.
Part. 1
x8 up.
Part. 0
Switch 0
x8 up.
Part. 0
Switch 1
x8 up.
Part. 0
Switch 2
x8 up.
Part. 0
Switch 3
x8 up.
Part. 1
Switch 4
SP VGA x1
Rear USB x1
SAS1 x8
NET0 x8
x8 up.
Part. 1
Debug Slot x1
MP
EB
Dotted line
devices
reside
on RIO
PCI-Express
Low Profile
Hot Plug Slots
Slot #
8 lanes
CPU #
10
11
12
13
14
15
16
x8
c0
x8
c0
x8
c0
x8
c0
x8
c0
x8
c0
x8
c1
x8
c1
x8
c1
x8
c1
x8
c1
x8
c1
x8
c1
x8
c1
x8
xx
x8
xx
T5-4/8 Native 2-Socket Configuration with One Root Domain Block fill color identifies Root Domain ownership
Block outline color identifies association to PM
Single non-redundant Domain
Switch 2 Slots crossed to maintain consistent Slot population order
Second level Switch 6 is partitioned
Slots drawn in order from left to right as in the actual chassis
differently from other configs
Native Config
Config
ID
PM1
CPU0
CPU1
CPU2
CPU3
MP
Front USB x1
Debug Slot x1
x8 up.
Part. 0
Switch 6
SAS0 x8
Switch 5
NET1 x8
x8 up.
Part. 0
x8 up.
Part. 0
x8 up.
Part. 1
x8 up.
Part. 0
Switch 0
x8 up.
Part. 3
x8 up.
Part. 0
Switch 1
x8 up.
Part. 3
x8 up.
Part. 0
Switch 2
x8 up.
Part. 3
x8 up.
Part. 1
Switch 3
Switch 4
SP VGA x1
Rear USB x1
SAS1 x8
NET0 x8
x8 up.
Part. 1
Debug Slot x1
MP
EB
Dotted line
devices
reside
on RIO
PCI-Express
Low Profile
Hot Plug Slots
Slot #
8 lanes
CPU #
10
11
12
13
14
15
16
x8
c0
x8
c0
x8
c0
x8
c0
x8
c2
x8
c2
x8
c2
x8
c2
x8
c1
x8
c1
x8
c1
x8
c1
x8
c3
x8
c3
x8
c3
x8
c3
T5-4/8 Native 4-Socket Configuration with One Root Domain Block fill color identifies Root Domain ownership
Block outline color identifies association to PM
Single non-redundant Domain
Switch 2 Slots crossed to maintain consistent Slot population order
Second level Switch 6 is partitioned
Slots drawn in order from left to right as in the actual chassis
differently from other configs
Native Config
Config
ID
PM1
CPU0
CPU1
PM2
PFM
PM3
PFM
CPU6
CPU7
MP
Front USB x1
Debug Slot x1
x8 up.
Part. 0
Switch 6
SAS0 x8
Switch 5
NET1 x8
x8 up.
Part. 0
x8 up.
Part. 0
x8 up.
Part. 1
x8 up.
Part. 0
Switch 0
x8 up.
Part. 3
x8 up.
Part. 0
Switch 1
x8 up.
Part. 3
x8 up.
Part. 0
Switch 2
x8 up.
Part. 3
x8 up.
Part. 1
Switch 3
Switch 4
SP VGA x1
Rear USB x1
SAS1 x8
NET0 x8
x8 up.
Part. 1
Debug Slot x1
MP
EB
Dotted line
devices
reside
on RIO
PCI-Express
Low Profile
Hot Plug Slots
Slot #
8 lanes
CPU #
10
11
12
13
14
15
16
x8
c0
x8
c0
x8
c0
x8
c0
x8
c6
x8
c6
x8
c6
x8
c6
x8
c1
x8
c1
x8
c1
x8
c1
x8
c7
x8
c7
x8
c7
x8
c7
Native Config
Config
ID
PM1
PM2
CPU0
CPU1
CPU2
CPU3
PM3
PFM
CPU6
CPU7
MP
Front USB x1
Debug Slot x1
x8 up.
Part. 0
Switch 6
SAS0 x8
Switch 5
NET1 x8
x8 up.
Part. 0
x8 up.
Part. 0
x8 up.
Part. 1
x8 up.
Part. 1
x8 up.
Part. 0
Switch 0
x8 up.
Part. 1
x8 up.
Part. 3
x8 up.
Part. 0
Switch 1
x8 up.
Part. 2
x8 up.
Part. 3
x8 up.
Part. 0
Switch 2
x8 up.
Part. 2
x8 up.
Part. 3
Switch 3
x8 up.
Part. 1
Switch 4
SP VGA x1
Rear USB x1
SAS1 x8
NET0 x8
x8 up.
Part. 1
Debug Slot x1
MP
EB
Dotted line
devices
reside
on RIO
PCI-Express
Low Profile
Hot Plug Slots
Slot #
8 lanes
CPU #
10
11
12
13
14
15
16
x8
c0
x8
c2
x8
c0
x8
c2
x8
c6
x8
c6
x8
c3
x8
c6
x8
c1
x8
c1
x8
c1
x8
c1
x8
c3
x8
c7
x8
c7
x8
c7
Native Config 0
Config
ID
PM1
PM2
PM3
CPU0
CPU1
CPU2
CPU3
CPU4
CPU5
CPU6
CPU7
MP
Front USB x1
Debug Slot x1
x8 up.
Part. 0
Switch 6
SAS0 x8
Switch 5
NET1 x8
x8 up.
Part. 0
x8 up.
Part. 0
x8 up.
Part. 1
x8 up.
Part. 1
x8 up.
Part. 0
Switch 0
x8 up.
Part. 1
x8 up.
Part. 2
x8 up.
Part. 3
x8 up.
Part. 0
Switch 1
x8 up.
Part. 1
x8 up.
Part. 2
x8 up.
Part. 3
x8 up.
Part. 0
Switch 2
x8 up.
Part. 1
x8 up.
Part. 2
x8 up.
Part. 3
Switch 3
x8 up.
Part. 0
x8 up.
Part. 1
Switch 4
SP VGA x1
Rear USB x1
SAS1 x8
NET0 x8
x8 up.
Part. 1
Debug Slot x1
MP
EB
Dotted line
devices
reside
on RIO
PCI-Express
Low Profile
Hot Plug Slots
Slot #
8 lanes
CPU #
10
11
12
13
14
15
16
x8
c0
x8
c2
x8
c0
x8
c2
x8
c4
x8
c6
x8
c4
x8
c6
x8
c1
x8
c3
x8
c1
x8
c3
x8
c5
x8
c7
x8
c5
x8
c7
Root Complex
PM
CPU
Switch
I/O Slot
/pci@300/pci@1/pci@0/pci@6
/pci@340/pci@1/pci@0/pci@6
/pci@380/pci@1/pci@0/pci@a
11
/pci@3c0/pci@1/pci@0/pci@e
/pci@400/pci@1/pci@0/pci@c
/pci@440/pci@1/pci@0/pci@c
10
/pci@480/pci@1/pci@0/pci@4
12
/pci@4c0/pci@1/pci@0/pci@8
/pci@500/pci@1/pci@0/pci@e
/pci@540/pci@1/pci@0/pci@e
13
/pci@580/pci@1/pci@0/pci@a
15
/pci@5c0/pci@1/pci@0/pci@8
/pci@600/pci@1/pci@0/pci@8
/pci@640/pci@1/pci@0/pci@8
14
/pci@680/pci@1/pci@0/pci@4
16
/pci@6c0/pci@1/pci@0/pci@4
All DIMMs must have the same Oracle Part Number on a PM (Processor
Module).
All PM's must be fully populated (32 DIMMs)
T5-4
(5U)
SATA SSDs
100GB
300GB
Disk LED's
Ready to
Remove
Fault
Status
SPARC T5 Processor
Features
16 S3 cores, 16-128 Strands
@ 3.6Ghz
Single or multi-threaded
operation per core
System scalability to 8
sockets
SPARC Core S3
ISA-based Cryptoacceleration
8MB Shared L3$
Integrated I/O
Downclock, Overclock
BoB
BoB
BoB
BoB
BoB
BoB
BoB
BoB
Memory
Control
Memory
Control
Memory
Control
Memory
Control
Coherence Unit
Coherence Unit
Coherence Unit
Coherence Unit
SPARC S3
Core
L3$ B0
L3$
B0
1MB,1
6-way
L3$ B2
1MB,1
6-way
L3$ B1
1MB,1
6-way
L3$ B3
1MB,1
6-way
L3$ B4
B0
L3$
B0
1MB,1
6-way
L3$ B6
1MB,1
6-way
L3$ B5
1MB,1
6-way
L3$ B7
1MB,1
6-way
FGU
Crypto
Coherency Links
12.8 Gbps per lane
- 12 lanes per link
Link 0
C0
C1
C2
C3
C4
C5
C6
C7
C8
C9
16 KB L1D$
IO
Subsystem
Link 1
128 KB L2$
16 KB L1I$
Link 2
Link 3
Link 4
Link 5
Link 6
T5 Processor Overview
SerDes
MI/O
SPAR SPAR
C
C
Core Core
SPAR SPAR
C
C
Core Core
MC MC
U
U
SPAR SPAR
C
C
Core Core
SPAR SPAR
C
C
Core Core
SerDes
Confidential Oracle Internal
providing 80 GB/s BW
SerDes
SerDes
16 S3 cores @ 3.6GHz
SerDes
scalability
Integrated 2x8 PCIe Gen 3
Advanced Power Management
with DVFS
2 Integer pipelines
Caches are all inclusive: L3 inclusive of L2; L2 inclusive of L1 (in this context,
inclusive refers to the fact a cached entry is always present in the next higher
level of cache)
Each core on SPARC T5 is capable of OoO execution, dual-issue of
instructions but in order commit.
Each core on SPARC T5 also includes cryptographic acceleration hardware,
accessible via user-level instructions.
S3 Core Recap
28nm port from 40nm T4
Out-of-order, dual-issue
High frequency achieved with 3.6GHz
16 stage integer pipeline
Dynamically threaded, one to eight strands
Accelerates 16 encryption algorithms and random number generation
IBM Power7
IBM Power7+
Intel Westmere/
Sandybridge
Operational Model
none
3 accelerators shared
across 8 cores
Userland
none
RSA, ECC
RSA, ECC
none
AES
AES
Message Digest /
Hash Functions
none
none
Supported
none
Supported
none
API Support
none
PKCS#11
Virtualization Support
Solaris Zones
Oracle VM for SPARC
none
??
Intel VT
T5 Servers
M6 Servers
S3 Core Overview
8-way threaded, dual-issue, OoO execution, in order commit
Dynamically threaded with hardware-optimized resource sharing
Support for Critical Threads
Deep pipeline for high frequency operation (3 GHz in 40 nm)
Balanced single-thread and multi-thread performance
5X better single-thread than SPARC T3 with equivalent multi-thread
performance
Enhanced instruction set to accelerate Oracle SW stack
PAUSE, fused compare-branch
Integrated user-level cryptographic acceleration
DES/3DES, AES, Kasumi, Camellia, MD5, SHA-1, SHA224/256/384/512,RSA, DSA, CRC32c
Foundation core for future technology / product nodes
52Copyright 2012, Oracle and/or its affiliates. All rights reserved.
Design Goals
Develop a common replacement core for T-series processors
Significantly improve the single strand performance of the T3 processor
-
Frequency
L1 Instruction Cache
L1 Data Cache
L2 Cache
# of Pipelines
# of Threads per Pipeline
Instructions per Thread
Out of Order Issue
Cryptography Acceleration
OVM for SPARC Compatible
SPARC V9 ISA Compatible
64KB
64KB
1
2
16KB
8KB
2
4
4 per cycle
Yes
None
No
Yes
1 per cycle
No
SPU
Yes
Yes
2 per cycle
Yes (36 instr window)
ISA Based
Yes
Yes
Applicability
Opportunity
Current Status
Logwriter, LMS
Up to 30% improvement in
efficiency
JAVA (JVM)
Compiler threads, GC Up to 2x improvement for app Support for JVM and JAVA
and priority mapping
apps to be CT aware is
startup,
Smooth
GC
support
integrated in JDK7U4
Coherence
Solaris
S11U1 / S10U11
Up to 20% improvement in
throughput
Integrated in Coherence
version 3.7.1 Patch 1
T5
T5
T5
T5
T5
T5
T
-
T5
T5
T5 System Interconnects
2-Way
Dual Socket
1-Way
Single Socket
4-Way
6-Way
8-Way
62Copyright 2012, Oracle and/or its affiliates. All rights reserved.
DIMMS
M5/T5
M5/T5
M5/T5
DIMMS
M5/T5
DIMMS
M5/T5
DIMMS
POINT-TO-POINT
LOCAL
INTERCONNECT
M5/T5
DIMMS
M5/T5
M5/T5
DIMMS
Latency for T5
T5-2
T5-4
Local Memory
136ns
Remote Memory
209ns
Cache to Cache
127ns
146ns
T5-8
155ns
Scalability of T4 vs T5
By numbers
Feature
Link Bandwidth
T4 Snoopy Based
Coherence Protocol
8 node snoops will consume 25% Link B/W &
increases linearly w/more nodes
Address serialization is done at Home Node. Home
Node broadcasts snoop request to all nodes. All
nodes except the requesting node require to
participate the snoop operation and provide snoop
response back to requesting node.
Message broadcast and response consume a lot of
link bandwidth.
T5 Directory Based
Coherence Protocol
8 node directory based will consume 5% of Link B/W
Address serialization is done at Directory Node.
Directory Node keeps track of which node hold each
cache line. Eliminating the need for broadcasting,
and relieve the L3$ from unnecessary foreign snoop
operation.
Directory filter the snoops sent to the share nodes.
Allow link bandwidth to be used more efficiently.
L3$ need to participate every snoop request from any Only the L3$ from the selective node require to
L3$ Performance other node. The L3$ performance can be dropped
participate the foreign snoop operation. L3$ has less
due to lots of foreign snoop requests.
distraction from foreign snoop request.
Easy to scale to large number of processor
Scalability
Limited to small scale of system.
environment.
Memory Controller of T4 vs T5
T4 MCU
T5 MCU
6.4 Gb/s
12.8 Gb/s
Not Supported
L0s, L1
Memory Buffer
Intel Milbrook2 MB
Advanced In-house MB
DDR3 Protocol
Burst length of 4
Burst length of 8
DDR3 Speed
800/1066
1066
DDR3 Device
1Gb/2Gb
2Gb/4Gb
RAS
Definition of Terms
Hot-plug:
refers to the fact that a component can be plugged and unplugged without
powering down the platform. It applies to both hot swap and hot service.
Hot service:
The system will notify the user when it is safe to remove the component.
End-to-End RAS
Built RAS from the inside out
Start with the processor, then memory,
system and IO, virtualization layer, and the
OS
Add Oracle Solaris Cluster software for
additional service availability
Fault Management Architecture (FMA)
binds all the layers together
Hypervisor
Enables software partitioning (LDoms)
virtualization and failure containment
Processor support for error clearing, correction
and collection
T5/M5 Processor
L1$ Tag, Status $ Data
Parity protection
Retry on error
L2$/L3$ Data
SEC/DED protection
Cache-line Sparing
L2$/L3$ Tags
SEC/DED protection
Inline Correction
Cache-line Sparing
L2$/L3$ Status & Directory
SEC/DED protection
Central Directory and Switch
Inline Correction
SEC/DED protection with in line correction
Architectural RegistersL2 Cache
Physical domain isolation
SEC/DED protection
CRC protected System Interconnect with message retry and
Precise Trap and
lane sparing
Hypervisor Correction and Retry
Deconfigurable directory chips, no loss of functionality,
minimized bandwidth loss
Unique to M5
Redundant Scalability Switch Boards
System
Redundant SPs with automatic failover
Redundant clock boards
Diagnosis to the FRU level on first fault
Power and Cooling
Advanced Power Management
Redundant hot-swap fans
Redundant hot-swap AC/DC
Dual grid power
System I/O
PCI-Express end-to-end CRC
PCI Express link retry
Hot-plug low profile PCI Express cards
Redundant, hot-plug boot disks
Alternate connections between M5 and IO
controllers
Memory
SDRAM Soft Errors
ECC Protection and Correction
Extended ECC Protection
4-bit Correction
Pin Steering
Channel Interconnect
CRC protection/Message Retry
Lane Sparing
T4-1/T4-2
T4-4
M4000/M5000
T5-2
T5-4/T5-8
M8000/M9000
M5-32
Dynamic Reconfiguration
X
X
Hot-plug PCIe
Memory Mirroring
X
X
Hard Domains
Common
Features
Logical Domains
Hot-plug disks
X
X
ECC Memory
Extended-ECC Memory
Instruction Retry
X
X
Partial ASIC Recovery
X
X
X
Fault Management
No ALOM Compatibility
ASR Support
FMA's Fault Proxy is used to keep ILOM's fault manager in sync with Solaris' fault
manager. Both will display the sum of all faults in the system.
Faults can be repaired from either side.
Fault Proxy communicates via the Ethernet Over USB connection.
IO faults are still diagnosed by Solaris.
For faults which diagnose resources as unusable, ILOM will add those resources to
the DDB. Resources excluded on next host reset.
When faults are repaired, ILOM automatically updates the DDB. Bringing
components back online requires a host reset.
Runs at SP boot. Tests devices on the SP FRU and its Ethernet port.
Status stored and converted to ereports after ILOM boots.
Fault proxy
SP
ereports
hostd
FETD
ip-transprt
LDC
Control Domain
ETM
ETM
faults
TCP/IP
ereports
LDC
IO Domain
ETM
faults
ip-transport
ETM
LDC
ETM
IO ereports are forwarded from the SP to the control domain, and then on to any
relevant IO domain
Faults are proxied between the SP, the control domain and any IO domains to provide
a single view of faults in the system.
Non-servicable faults such as memory faults are not proxied.
The SP and the control domain can view and manage all faults in the system.
An IO domain can only view and manage faults local to the domain.
ASR Support