Base Standard Capabilities
Capability / Feature | Description | ||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Chassis | Two rack unit (2RU) chassis | ||||||||||||||||||||||||||||
CPU | One or two Intel Xeon E5-2600 v3 or v4 series processor family CPUs | ||||||||||||||||||||||||||||
Chipset | Intel® C610 series chipset | ||||||||||||||||||||||||||||
Memory | 24 slots for registered ECC DIMMs (RDIMMs) or load-reduced DIMMs (LRDIMMs) | ||||||||||||||||||||||||||||
Multi-bit Error Protection | This server supports multi-bit error protection. | ||||||||||||||||||||||||||||
Expansion slots | Up to six PCIe slots (on two riser cards)
| ||||||||||||||||||||||||||||
Video | The Cisco Integrated Management Controller (CIMC) provides video using the Matrox G200e video/graphics controller:
| ||||||||||||||||||||||||||||
Internal storage devices | Drives are installed into front-panel drive bays that provide hot-pluggable access.
| ||||||||||||||||||||||||||||
Cisco Flexible Flash drives | The server supports up to two internal 32 GB or two internal 64 GB Cisco Flexible Flash drives (SD cards). The second SD card is blank and can be used to mirror the first SD card. It can be used to protect the Hypervisor Partition with RAID1. | ||||||||||||||||||||||||||||
Interfaces |
| ||||||||||||||||||||||||||||
Power subsystem | Up to two of the following hot-swappable power supplies:
| ||||||||||||||||||||||||||||
Storage controller | Cisco 12G SAS Modular RAID controller card with internal SAS connectivity.
Can be purchased alone, or along with an onboard Flash-Backed Write Cache (FBWC) upgrade option, as shown in the table below
All versions of the UCSC-MRAID12G RAID controller support up to 24 internal SAS (limited to 12 drives for this server). Cisco 12 Gbps Modular SAS HBA with internal SAS connectivity
Cisco 9300-8E 12G SAS HBA with external SAS connectivity
| ||||||||||||||||||||||||||||
Embedded NIC | Two embedded (on the motherboard) Intel i350 GbE ports, supporting the following:
| ||||||||||||||||||||||||||||
Modular LAN on Motherboard (mLOM) slot | The mLOM slot can flexibly accommodate the following cards:
| ||||||||||||||||||||||||||||
WoL | The 1-Gb Base-T Ethernet LAN ports support the wake-on-LAN (WoL) standard. | ||||||||||||||||||||||||||||
Front Panel | A front panel controller provides status indications and control buttons | ||||||||||||||||||||||||||||
ACPI | This server supports the advanced configuration and power interface (ACPI) 4.0 standard. | ||||||||||||||||||||||||||||
Fans | Chassis: 6 hot-swappable fans for front-to-rear cooling | ||||||||||||||||||||||||||||
Integrated management processor | Baseboard Management Controller (BMC) running Cisco Integrated Management Controller (CIMC) firmware. Depending on your CIMC settings, the CIMC can be accessed through the 1-GbE dedicated management port, the 1-GbE LOM ports, or a Cisco virtual interface card (VIC). | ||||||||||||||||||||||||||||
Boot drives | Up to two optional SATA drives can be installed internal to the chassis on riser 1. The two SATA boot drives are supported only on the 24-drive backplane chassis version and are managed in AHCI mode, using OS-based software RAID. | ||||||||||||||||||||||||||||
Footer | |||||||||||||||||||||||||||||
The four Intel i350 ports are provided on an optional card that plugs into the mLOM slot, and are separate from the two embedded (on the motherboard) LAN ports | |||||||||||||||||||||||||||||
NCSI = Network Communications Services Interface protocol. An NCSI slot is powered even when the server is in standby power mode. | |||||||||||||||||||||||||||||
The drives are available as both self-encrypted drives (SED) and non self-encrypted drives. |
Available Riser Cards
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
Riser 1 Options | ||||||||||
UCSC-PCI-1A-240M4 | C240 M4 PCIe Riser 1 Assy (option A) (2 PCIe slots: 1x8 and 1x16 GPU capable) | |||||||||
UCSC-PCI-1B-240M4 | C240 M4 PCIe Riser 1 Assy (option B) (3 PCIe slots: 3x8) | |||||||||
UCSC-PCI-1C-240M4 | C240 M4 PCIe Riser 1 Assy (option C) (2 PCIe slots: 1x8 and 1x16 plus connectors for 2 SATA boot drives) | |||||||||
Riser 2 Options | ||||||||||
UCSC-PCI-2-C240M4 | PCIe Riser Board (Riser 2) for C240 M4 (3 slots: 2x8 and 1x16) | |||||||||
Footnote: | ||||||||||
[1] | Riser card 1 is the one on the left when viewed from the back of the server and riser card 2 is on the right. | |||||||||
[2] | The selection of riser card 1 determines the number and type of PCIe cards and SATA boot drives supported in the riser. | |||||||||
[3] | If no riser is selected, a riser blanking panel will be installed. You will not be able to install any PCIe cards without a riser selected |
Available CPU's
Product ID (PID) | Intel Number | Clock Freq (GHz) | Power (W) | Cache Size (MB) | Cores | QPI | Highest DDR4 DIMM Clock Support (MHz) [1] | |||
---|---|---|---|---|---|---|---|---|---|---|
E5-2600 v4 Series Processor Family CPUs | ||||||||||
UCS-CPU-E52699E | E5-2699 v4 | 2.2 | 145 | 55 | 22 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52698E | E5-2698 v4 | 2.2 | 135 | 50 | 20 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52697AE | E5-2697A v4 | 2.6 | 145 | 40 | 16 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52697E | E5-2697 v4 | 2.3 | 145 | 45 | 18 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52695E | E5-2695 v4 | 2.1 | 120 | 45 | 18 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52690E | E5-2690 v4 | 2.6 | 135 | 35 | 14 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52683E | E5-2683 v4 | 2.1 | 120 | 40 | 16 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52680E | E5-2680 v4 | 2.4 | 120 | 35 | 14 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52667E | E5-2667 v4 | 3.2 | 135 | 25 | 8 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52660E | E5-2660 v4 | 2 | 105 | 35 | 14 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52650E | E5-2650 v4 | 2.2 | 105 | 30 | 12 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52650LE | E5-2650L v4 | 1.7 | 65 | 35 | 14 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52643E | E5-2643 v4 | 3.4 | 135 | 20 | 6 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52640E | E5-2640 v4 | 2.4 | 90 | 25 | 10 | 8 GT/s | 2133 | |||
UCS-CPU-E52637E | E5-2637 v4 | 3.5 | 135 | 15 | 4 | 9.6 GT/s | 2400 | |||
UCS-CPU-E52630E | E5-2630 v4 | 2.2 | 85 | 25 | 10 | 8 GT/s | 2133 | |||
UCS-CPU-E52630LE | E5-2630L v4 | 1.8 | 55 | 25 | 10 | 8 GT/s | 2133 | |||
UCS-CPU-E52623E | E5-2623 v4 | 2.6 | 85 | 10 | 4 | 8 GT/s | 2133 | |||
UCS-CPU-E52620E | E5-2620 v4 | 2.1 | 85 | 20 | 8 | 8 GT/s | 2133 | |||
UCS-CPU-E52609E | E5-2609 v4 | 1.7 | 85 | 20 | 8 | 6.4 GT/s | 1866 | |||
UCS-CPU-E52658E | E5-2658 v4 | 2.3 | 105 | 35 | 14 | 9.6 GT/s | 2400 | |||
E5-2600 v3 Series Processor Family CPUs | ||||||||||
UCS-CPU-E52699D | E5-2699 v3 | 2.3 | 145 | 45 | 18 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52698D | E5-2698 v3 | 2.3 | 135 | 40 | 16 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52697D | E5-2697 v3 | 2.6 | 145 | 35 | 14 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52695D | E5-2695 v3 | 2.3 | 120 | 35 | 14 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52690D | E5-2690 v3 | 2.6 | 135 | 30 | 12 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52683D | E5-2683 v3 | 2 | 120 | 35 | 14 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52680D | E5-2680 v3 | 2.5 | 120 | 30 | 12 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52670D | E5-2670 v3 | 2.3 | 120 | 30 | 12 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52667D | E5-2667 v3 | 3.2 | 135 | 20 | 8 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52660D | E5-2660 v3 | 2.6 | 105 | 25 | 10 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52658D | E5-2658 v3 | 2.2 | 105 | 30 | 12 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52650D | E5-2650 v3 | 2.3 | 105 | 25 | 10 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52650LD | E5-2650L v3 | 1.8 | 65 | 30 | 12 | 9.6 GT/s | 1866 | |||
UCS-CPU-E52643D | E5-2643 v3 | 3.4 | 135 | 20 | 6 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52640D | E5-2640 v3 | 2.6 | 90 | 20 | 8 | 8 GT/s | 1866 | |||
UCS-CPU-E52637D | E5-2637 v3 | 3.5 | 135 | 15 | 4 | 9.6 GT/s | 2133 | |||
UCS-CPU-E52630D | E5-2630 v3 | 2.4 | 85 | 20 | 8 | 8 GT/s | 1866 | |||
UCS-CPU-E52630LD | E5-2630L v3 | 1.8 | 55 | 20 | 8 | 8 GT/s | 1866 | |||
UCS-CPU-E52623D | E5-2623 v3 | 3 | 105 | 10 | 4 | 8 GT/s | 1866 | |||
UCS-CPU-E52620D | E5-2620 v3 | 2.4 | 85 | 15 | 6 | 8 GT/s | 1866 | |||
UCS-CPU-E52609D [2] | E5-2609 v3 | 1.9 | 85 | 15 | 6 | 6.4 GT/s | 1600 | |||
Footnote: | ||||||||||
[1] | If higher or lower speed DIMMs are selected than what is shown in the table for a given CPU, the DIMMs will be clocked at the lowest common denominator of CPU clock and DIMM clock | |||||||||
[Caveats] | - | |||||||||
[1] | You can select either one processor or two identical processors. | |||||||||
[2] | For optimal performance, select DIMMs with the highest clock speed for a given processor | |||||||||
[Specs] | - | |||||||||
[1] | Intel Xeon E5-2600 v3 and v4 series processor family CPUs | |||||||||
[2] | Intel C610 series chipset | |||||||||
[3] | Cache size of up to 55 MB |
Available DIMMs and Memory Mirroring
Product ID (PID) | PID Description | Voltage | Ranks /DIMM | |||||||
---|---|---|---|---|---|---|---|---|---|---|
2400-MHz DIMM Options | ||||||||||
UCS-ML-1X644RV-A | 64GB DDR4-2400-MHz LRDIMM/PC4-19200/quad rank/x4 | 1.2 V | 4 | |||||||
UCS-MR-1X322RV-A | 32GB DDR4-2400-MHz RDIMM/PC4-19200/dual rank/x4 | 1.2 V | 2 | |||||||
UCS-MR-1X161RV-A | 16GB DDR4-2400-MHz RDIMM/PC4-19200/single rank/x4 | 1.2 V | 1 | |||||||
UCS-MR-1X081RV-A | 8 GB DDR4-2400-MHz RDIMM/PC4-19200/single rank/x4 | 1.2 V | 1 | |||||||
UCS-ML-1X324RV-A | 32GB DDR4-2400-MHz LRDIMM/PC4-19200/quad rank/x4 | 1.2 V | 4 | |||||||
UCS-MR-1X162RV-A | 16GB DDR4-2400-MHz RDIMM/PC4-19200/dual rank/x4 | 1.2 V | 2 | |||||||
2133-MHz DIMM Options | ||||||||||
UCS-MR-1X648RU-A [1], [2] | 64GB DDR4-2133-MHz TSV-RDIMM/PC4-17000/octal rank/x4 | 1.2 V | 8 | |||||||
UCS-MR-1X322RU-A | 32GB DDR4-2133-MHz RDIMM/PC4-17000/dual rank/x4 | 1.2 V | 2 | |||||||
UCS-ML-1X324RU-A | 32GB DDR4-2133-MHz LRDIMM/PC4-17000/quad rank/x4 | 1.2 V | 4 | |||||||
UCS-MR-1X162RU-A | 16GB DDR4-2133-MHz RDIMM/PC4-17000/dual rank/x4 | 1.2 V | 2 | |||||||
UCS-MR-1X081RU-A | 8GB DDR4-2133-MHz RDIMM/PC4-17000/single rank/x4 | 1.2 V | 1 | |||||||
Memory Mirroring Option | ||||||||||
N01-MMIRROR | Memory mirroring option | - | - | |||||||
Footnote: | ||||||||||
[1] | Power capping is not supported when using 64GB TSV-RDIMMS. | |||||||||
[2] | NVIDIA GPUs can support only less than 1 TB of total memory in the server. Do not install more than fourteen 64-GB DIMMs when using an NVIDIA GPU card in this server. | |||||||||
[1-CPU] | Without memory mirroring: Select from 1 to 12 DIMMs. | |||||||||
[2-CPU] | Without memory mirroring: Select from 1 to 12 DIMMs. | |||||||||
[1] | System performance is optimized when the DIMM type and quantity are equal for both CPUs, and when all channels are filled equally across the CPUs in the server. | |||||||||
[2] | System speed is dependent on how many DIMMs are populated per channel and the CPU DIMM speed support. | |||||||||
[3] | When you mix 2400 MHz RDIMMs of different sizes (for example, 4x32GB + 4x16GB), the memory speed clocks one step down to 2133 MHz. | |||||||||
[4] | Do not mix RDIMMs with LRDIMMs | |||||||||
[5] | Do not mix 64GB DDR4-2133-MHz TSV-RDIMMs with any other DIMMs | |||||||||
[6] | Single-rank DIMMs can be mixed with dual-rank DIMMs in the same channel | |||||||||
[7] | Do not mix quad-rank DIMMs with single- or dual-rank DIMMs in the same channel | |||||||||
[8] | DIMMs for CPU 1 and CPU 2 (when populated) must always be configured identically. | |||||||||
[9] | When using mirroring, DIMMs must be installed in identical pairs across paired DDR4 buses. That is, mirrored pairs in channels A and B must be identical and pairs in channels C and D must be identical. However, the DIMMs used in channels A and B can be different from those in channels C and D. | |||||||||
[10] | Memory mirroring reduces the amount of available memory by 50% (quantity of DIMMs must be even for mirroring). | |||||||||
[11] | Non-ECC DIMMs are not supported. | |||||||||
[12] | Cisco memory from previous generation servers (DDR3) is not compatible with this server |
Available RAID Controllers
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
Controllers for Internal Drives | ||||||||||
- | Note that if the following Cisco 12G SAS Modular RAID controller or Cisco 12Gbps Modular SAS HBA controller is selected, it is factory-installed in the dedicated internal slot. | |||||||||
UCSC-MRAID12G | Cisco 12G SAS Modular RAID Controller
| |||||||||
UCSC-SAS12GHBA | Cisco 12 Gbps Modular SAS HBA
| |||||||||
UCSC-PSAS12GHBA | Same functionality as UCSC-SAS12GHBA, but vendor is Microsemi | |||||||||
Flash-Backed Write Cache (FBWC) Upgrade Options for Cisco 12G SAS Modular RAID controller | ||||||||||
UCSC-MRAID12G-1GB | 1 GB FBWC, which includes a 1 GB MB memory plus a SuperCap for write cache backup. Supports JBOD, RAID 0, 1, 10, 5, 6, 50, and 60. | |||||||||
UCSC-MRAID12G-2GB | 2 GB FBWC, which includes a 2 GB MB memory plus a SuperCap for write cache backup. Supports JBOD, RAID 0, 1, 10, 5, 6, 50, and 60. | |||||||||
UCSC-MRAID12G-4GB | 4 GB FBWC, which includes a 4 GB MB memory plus a SuperCap for write cache backup. Supports JBOD, RAID 0, 1, 10, 5, 6, 50, and 60. | |||||||||
SAS HBA for External JBOD Attach Two SFF8644 mini-SAS HD connectors on this card are accessible at the rear of the chassis. | ||||||||||
UCSC-SAS9300-8E | Cisco 9300-8e 12G SAS HBA for external JBOD attach Supports external JBOD using X8 wide SAS ports. | |||||||||
RAID Configuration Options (not available for Cisco 12 Gbps Modular SAS HBA or embedded software RAID) | ||||||||||
R2XX-SRAID0 | Enable Single Disk Raid 0 Setting | |||||||||
R2XX-RAID0 | Factory preconfigured RAID striping option Enable RAID 0 Setting. Requires a minimum of one hard drive. | |||||||||
R2XX-RAID1 | Factory preconfigured RAID mirroring option Enable RAID 1 Setting. Requires exactly two drives with the same size, speed, capacity. | |||||||||
R2XX-RAID5 | Factory preconfigured RAID option Enable RAID 5 Setting. Requires a minimum of three drives of the same size, speed, capacity. | |||||||||
R2XX-RAID6 | Factory preconfigured RAID option Enable RAID 6 Setting. Requires a minimum of four drives of the same size, speed, capacity. | |||||||||
Footnote: | ||||||||||
[Caveats] | - | |||||||||
[1] | A maximum of one Cisco 9300-8e 12G SAS HBA can be installed, and it can be installed in any slot (depending on the number of CPUs installed). The system can also support combinations of one Cisco 9300-8e 12G SAS HBA and a Storage Accelerator card along with either a Cisco 12G SAS Modular RAID Controller, or a Cisco 12 Gbps Modular SAS HBA | |||||||||
[2] | For the Cisco 12G SAS Modular RAID controller, you can choose an optional RAID configuration (up to RAID 0, 1, 10, 5, 6, 50, 60 if optional FBWC option is chosen), which is preconfigured at the factory. The RAID level you choose must be an available RAID choice for the controller selected. RAID levels 50 and 60 are supported, although they are not available as configuration options. It can also be combined with AHCI support for internal SSDs (SATA Boot Drives). | |||||||||
[Approved Configurations] | - | |||||||||
[1] | The Cisco 12G SAS Modular RAID controller option supports up to 12 internal SAS HDDs with up to RAID 0, 1, 10, 5, 6, 50, 60 support (with FBWC option chosen). | |||||||||
[2] | The Cisco 12 Gbps Modular SAS HBA option supports up to 12 internal SAS HDDs with JBOD support. | |||||||||
[3] | The Cisco 9300-8e 12G SAS HBA supports up to 8 external SAS ports with JBOD support. | |||||||||
[SAS HBA (internal HDD/SSD/JBOD support)] | - | |||||||||
[-] | Instead of a RAID controller, you can choose a SAS HBA for internal drive connectivity
(non-RAID):
Cisco 12 Gbps Modular SAS HBA, which plugs into a dedicated RAID controller slot. | |||||||||
[SAS HBA (external JBOD support)] | - | |||||||||
[-] | In addition to a RAID controller or JBOD SAS HBA for internal drives, you can choose the
following SAS HBA for external drive connectivity (non-RAID):
Cisco 9300-8e 12G SAS HBA (provides 8 SAS ports for external JBOD connectivity). | |||||||||
[RAID Volumes and Groups] | - | |||||||||
[] | When creating each RAID volume, follow these guidelines:
|
Available 3.5"" LFF Hot-pluggable Drives
Product ID (PID) | PID Description | Drive Type | Capacity | |||||||
---|---|---|---|---|---|---|---|---|---|---|
HDDs | ||||||||||
UCS-HD600G15CK9 | 600 GB 12G SAS 15K RPM LFF HDD (SED) | SAS | 600 GB | |||||||
UCS-HD6T12GAK9 | 6 TB 7.2K RPM HDD (4K sector format, SED) | SAS | 6 TB | |||||||
UCS-HD4T12GK9 | 4 TB 7.2K RPM HDD (SED) | SAS | 4 TB | |||||||
UCS-HD10T7KL4K [2] | 10 TB 12G SAS 7.2K RPM LFF HDD (4K sector format) | SAS | 10 TB | |||||||
UCS-HD10T7KEM | 10 TB 12G SAS 7.2K RPM LFF HDD (512e) | SAS | 10 TB | |||||||
UCS-HD8T7KEM | 8 TB 12G SAS 7.2K RPM LFF HDD (512e) | SAS | 8 TB | |||||||
UCS-HD6T7KEM | 6 TB 12G SAS 7.2K RPM LFF HDD (512e) | SAS | 6 TB | |||||||
UCS-HD1T7KL12G | 1 TB 12G SAS 7.2K RPM LFF HDD | SAS | 6 TB | |||||||
UCS-HD4T7KL12G | 4 TB 12G SAS 7.2K RPM LFF HDD | SAS | 4 TB | |||||||
UCS-HD2T7KL12G | 2 TB 12G SAS 7.2K RPM LFF HDD | SAS | 2 TB | |||||||
UCS-HD8T7KL4K [2] | 8 TB 12G SAS 7.2K RPM LFF HDD (4K sector format) | SAS | 8 TB | |||||||
UCS-HD6T7KL4K [2] | 6 TB 12G SAS 7.2K RPM LFF HDD (4K sector format) | SAS | 6 TB | |||||||
UCS-HD4T7KL4K [2] | 4 TB 12G SAS 7.2K RPM LFF HDD (4K sector format) | SAS | 4 TB | |||||||
UCS-HD2T7KL4K [2] | 2 TB 12G SAS 7.2K RPM LFF HDD (4K sector format) | SAS | 2 TB | |||||||
UCS-HD12TB10KHY-E | 1.2 TB 3.5 inch 12G SAS 10K RPM HDD | SAS | 1.2 TB | |||||||
UCS-HDD1TI2F212 | 1 TB SAS 7.2K RPM LFF HDD | SAS | 1 TB | |||||||
UCS-HD600G15KHY-E | 600 GB 3.5 inch Hybrid 6G SAS 15K RPM HDD | SAS | 600 GB | |||||||
UCS-HD300G15KHY-E | 300 GB 3.5 inch Hybrid 6G SAS 15K RPM HDD | SAS | 300 GB | |||||||
SSDs | ||||||||||
UCS-SD800GBCK9 | 800 GB 2.5"" Enterprise performance SAS SSD (10X FWPD-SED) | SAS | 800 GB | |||||||
UCS-SD400GBCK9 | 400 GB 2.5"" Enterprise performance SAS SSD (10X FWPF-SED) | SAS | 400 GB | |||||||
UCS-SD16TG0KHY-EP | 1.6 TB 3.5 inch Enterprise performance 6G SATA SSD (3X FWPD, Intel 3610) | SATA | 1.6 TB | |||||||
UCS-SD800G0KHY-EP | 800 GB 3.5 inch Hybrid Enterprise Performance 12G SSD (10X FWPD, Samsung 1635) | SAS | 800 GB | |||||||
UCS-SD480G0KHY-EP | 480 GB 3.5 inch Enterprise Performance 6G SATA SSD (3X FWPD, Intel 3610) | SATA | 480 GB | |||||||
UCS-SD120G0KHY-EV | 120 GB 3.5 inch Enterprise Performance 6G SATA SSD (1FWPD) | SATA | 120 GB | |||||||
Boot Drives | ||||||||||
UCS-SD120GBE1K9 | 120GB Enterprise Value SSD (SATA) (1X FWPD, SED) (boot) | SATA | 120 GB | |||||||
UCS-SD16TBKS4-EB | 1.6 TB 2.5 inch Enterprise Value 6G SATA SSD (BOOT) | SATA | 1.6 TB | |||||||
UCS-SD960G0KSB-EV | 960 GB 2.5 inch Enterprise Value 6G SATA SSD (BOOT) | SATA | 960 GB | |||||||
UCS-SD480G0KSB-EV | 480 GB 2.5 inch Enterprise Value 6G SATA SSD (BOOT) | SATA | 480 GB | |||||||
UCS-SD480GBKS4-EB | 480 GB 2.5 inch Enterprise Value 6G SATA SSD (Boot) | SATA | 480 GB | |||||||
UCS-SD240GBKS4-EB | 240 GB 2.5 inch Enterprise Value 6G SATA SSD (boot) | SATA | 240 GB | |||||||
UCS-SD120GBKS4-EB | 120 GB 2.5 inch Enterprise Value 6G SATA SSD (boot) | SATA | 120 GB | |||||||
UCS-SD120G0KSB-EV | 120 GB 2.5 inch Enterprise Value 6G SATA SSD (BOOT) | SATA | 120 GB | |||||||
Footnote: | ||||||||||
[Notes] | - | |||||||||
[1] | For all self-encrypting drives (SED), standalone Management (CIMC) is supported for configuring and managing local keys (UCSM is not supported). Also, 4K format drives do not support Vmware and require UEFI boot. | |||||||||
[2] | UCS Rack Servers require minimum firmware version 2.0(4) to support 4K sector size drives. | |||||||||
[3] | EFI Mode must be used when booting from 4K sector size drives, legacy mode is not supported. Ensure that 4K sector size and 512 byte sector size drives are not mixed in the same RAID volume. | |||||||||
[Approved Configurations] | - | |||||||||
[1] | Cisco 12G SAS Modular RAID Controller
| |||||||||
[2] | Cisco 12 Gbps Modular SAS HBA
If you selected a Cisco 12 Gbps Modular SAS HBA, select up to 12 SAS HDDs. | |||||||||
[3] | Systems Using Boot Drives f you are configuring a system that uses SATA SSD boot drives, choose up to two identical boot drives | |||||||||
[Caveats] | - | |||||||||
[1] | You can choose only SAS HDDs when using the Cisco 12G SAS Modular RAID Controller or Cisco 12 Gbps Modular SAS HBA. | |||||||||
[2] | If you order one or two boot drives, you can order a maximum of one GPU and it must be installed in Riser 2 (UCSC-PCI-2-C240M4) slot 5. | |||||||||
[3] | A maximum of two of the 400 GB 3.5-inch hybrid enterprise performance 12G SSDs can be installed. |
Available PCIe Option Cards
Product ID (PID) | PID Description | Card Height | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
Modular LAN on Motherboard (mLOM) | ||||||||||
UCSC-MLOM-C10T-02 | Cisco UCS VIC1227T VIC MLOM - Dual Port 10GBaseT | N/A | ||||||||
UCSC-MLOM-CSC-02 | Cisco UCS VIC1227 VIC MLOM - Dual Port 10Gb SFP+ | N/A | ||||||||
UCSC-MLOM-IRJ45 | Intel i350 quad-port MLOM NIC | N/A | ||||||||
UCSC-MLOM-C40Q-03 | Cisco VIC 1387 Dual Port 40Gb QSFP CNA MLOM | N/A | ||||||||
Virtual Interface Cards (VICs) | ||||||||||
UCSC-PCIE-CSC-02 | Cisco VIC 1225 Dual Port 10Gb SFP+ CNA | Half | ||||||||
UCSC-PCIE-C10T-02 | Cisco VIC 1225T Dual Port 10GBaseT CNA | Half | ||||||||
UCSC-PCIE-C40Q-03 | Cisco VIC 1385 Dual Port 40Gb QSFP+ CNA w/RDMA | Half | ||||||||
Network Interface Cards (NICs) | ||||||||||
1 Gb NICs | ||||||||||
N2XX-ABPCI03-M3 | Broadcom 5709 Quad Port 1Gb w/TOE iSCSI for M3 Servers | Half | ||||||||
N2XX-ABPCI01-M3 | Broadcom 5709 Dual-Port Ethernet PCIe Adapter M3 & later | Half | ||||||||
UCSC-PCIE-IRJ45 | Intel i350 Quad Port 1Gb Adapter | Half | ||||||||
10 Gb NICs | ||||||||||
N2XX-AIPCI01 | Intel X520 Dual Port 10Gb SFP+ Adapter | Half | ||||||||
UCSC-PCIE-ITG | Intel X540 Dual Port 10GBase-T Adapter | Half | ||||||||
UCSC-PCIE-QNICBT | QLogic QLE8442 dual-port 10Gbase-T NIC | Half | ||||||||
UCSC-PCIE-QNICSFP | QLogic QLE8442 dual-port 10G SFP+ NIC | Half | ||||||||
UCSC-PCIE-ID10GC | Intel X550-T2 dual-port 10GBase-T NIC | Half | ||||||||
UCSC-PCIE-ID10GF | Intel X710-DA2 dual-port 10G SFP+ NIC | Half | ||||||||
UCSC-PCIE-IQ10GF | Intel X710 quad-port 10G SFP+ NIC | Half | ||||||||
UCSC-PCIE-ID40GF | Intel XL710 dual-port 40G QSFP+ NIC | Half | ||||||||
10 Gb Converged Network Adapters (CNAs) | ||||||||||
UCSC-PCIE-E14102B | Emulex OCe14102B-UX dual-port 10G SFP+ with iSCSI, FCoE CNA | Half | ||||||||
UCSC-PCIE-E14102 | Emulex OCe14102-UX dual-port 10 GbE FCoE CNA | Half | ||||||||
UCSC-PCIE-Q8362 | Qlogic QLE8362 dual-port 10 GbE FCoE CNA | Half | ||||||||
Host Bus Adapters (HBAs) | ||||||||||
N2XX-AQPCI05 | Qlogic QLE2562 Dual Port 8Gb Fibre Channel HBA | Half | ||||||||
UCSC-PCIE-Q2672 | Qlogic QLE2672-CSC, 16Gb Fibre Channel HBA with SR Optics N2XX-AEPCI05 Emulex LPe 12002 Dual Port 8Gb Fibre Channel HBA | Half | ||||||||
N2XX-AEPCI05 | Emulex LPe 12002 Dual Port 8Gb Fibre Channel HBA | Half | ||||||||
UCSC-PCIE-E16002 | Emulex LPe16002-M6, 16Gb Fibre Channel HBA with SR Optics | Half | ||||||||
UCS Storage Accelerators [3] | ||||||||||
UCSC-F-S13002 | UCS Rack PCIe Storage 1300 GB SanDisk SX350 Medium Endurance | Half | ||||||||
UCSC-F-S16002 | UCS Rack PCIe Storage 1600 GB SanDisk SX350 Medium Endurance | Half | ||||||||
UCSC-F-S32002 | UCS Rack PCIe Storage 3200 GB SanDisk SX350 Medium Endurance | Half | ||||||||
UCSC-F-S64002 | UCS Rack PCIe Storage 6400 GB SanDisk SX350 Medium Endurance | Full | ||||||||
UCSC-F-FIO-5200PS | UCS 5200 GB Fusion ioMemory3 PX Performance line for C-Series | Full | ||||||||
UCSC-F-FIO-2600PS | UCS 2600 GB Fusion ioMemory3 PX Performance line for C-Series | Half | ||||||||
UCSC-F-FIO-1300PS | UCS 1300 GB Fusion ioMemory3 PX Performance line for C-Series | Half | ||||||||
UCSC-F-FIO-1000PS | UCS 1000 GB Fusion ioMemory3 PX Performance line for C-Series | Half | ||||||||
UCS NVMe/PCie Add in Cards | ||||||||||
UCSC-F-H38001 | 3.8 TB NVMe/PCIe storage (HGST SN150) Read Intensive | Half | ||||||||
UCSC-F-I20003 | 2.0 TB NVMe/PCIe storage (Intel P3600) High Endurance | Half | ||||||||
UCSC-F-H19001 | 1.9 GB NVMe/PCIe storage (HGST SN150) Read Intensive | Half | ||||||||
UCSC-F-I160010 | 1.6 TB NVMe/PCIe storage (Intel P3700) High Endurance | Half | ||||||||
UCSC-F-I12003 | 1.2 GB NVMe/PCIe storage (Intel P3600) High Endurance | Half | ||||||||
UCSC-F-I80010 | 800 GB NVMe/PCIe storage (Intel P3700) High Endurance | Half | ||||||||
Footnote: | ||||||||||
[1] | The GPU cards must be inserted into a full length x16 (electrical) PCIe slot. The rest of the PCIe cards are x8 (electrical) and can be plugged into any PCIe slot of riser 1 or riser 2. | |||||||||
[2] | The mLOM cards do not plug into any of the riser 1 or riser 2 card slots | |||||||||
[4] | A maximum of six storage accelerator cards are supported and some are riser dependent. 1.3 TB, 2.6 TB, or 3.2 TB cards should not be installed in slots 5 or 6 so the total number of those capacity points is four. The Cisco 9300-8e 12G SAS HBA also can only be installed in riser 1 | |||||||||
[Caveats] | - | |||||||||
[-] | If you choose an external drives controller (Cisco 9300-8e 12G SAS HBA), it will consume one PCIe slot. | |||||||||
[-] | A maximum of six storage accelerator cards are supported and some are riser dependent. 1.3 TB, 2.6 TB, or 3.2 TB cards should not be installed in to slots 5 or 6 so the total number of those capacity points is four. | |||||||||
[For 1-CPU systems:] | Only the PCIe slots on PCIe riser 1 are available for 1-CPU system. | |||||||||
[-] | The PCIe slots on riser 2 are not supported on 1-CPU systems. The slots are full-height PCIe slots 4, 5, and 6 | |||||||||
[-] | Only a single plug-in PCIe VIC card may be installed on a 1-CPU system, and it must be installed in slot 2 of riser 1. You can order an mLOM VIC card to be installed in the mLOM slot internal to the chassis and thus have two VIC cards in operation at the same time. | |||||||||
[For 2-CPU systems:] | Depending on the riser 1 option chosen, up to six PCIe slots are available, three on PCIe riser 1 (PCIe slots 1, 2, and 3) and three on PCIe riser 2 (PCIe slots 4, 5, and 6). | |||||||||
[-] | All of the slots are full-height. | |||||||||
[-] | Two plug-in PCIe VIC cards can be installed in 2-CPU systems, using slots 2 and 5. In addition, you can order an mLOM VIC card, which is installed in the mLOM slot inside the chassis and thus have three VIC cards in operation at the same time. | |||||||||
[-] | If GPUs are installed in slots 2 (Riser 1 option A) and 5 (Riser 2), the NCSI capability automatically switches over to slots 1 (Riser 1 option A) and 4 (Riser 2). Therefore, Cisco PCIe VICs can be installed in slots 1 and 4 if GPUs are installed in slots 2 and 5. | |||||||||
[Notes] | UCSM managed servers are discoverable only if a VIC 1225 is installed in slot 2 or a VIC 1227/1227T is installed in the MLOM slot. If you install two GPUs, they must be located in slots 2 and 5. Therefore, if two GPUs are installed, UCSM managed servers are discoverable only if you install a VIC 1227/1227T in the MLOM slot. | |||||||||
[-] | The quantity and type of PCIe cards that can be installed depends in the riser card options. |
Available Network Card Accessories (Optional)
Product ID(PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
Available Twinax Cables | ||||||||||
SFP-H10GB-CU1M | 10GBASE-CU SFP+ Cable (1 M) | |||||||||
SFP-H10GB-CU3M | 10GBASE-CU SFP+ Cable (3 M) | |||||||||
SFP-H10GB-CU5M | 10GBASE-CU SFP+ Cable (5 M) | |||||||||
SFP-H10GB-ACU7M | 10GBASE-CU SFP+ Cable (7 M) | |||||||||
SFP-H10GB-ACU10M | 10GBASE-CU SFP+ Cable (10 M) | |||||||||
Available SFP Modules | ||||||||||
SFP-10G-SR | 10GBASE-SR SFP+ Module 850 nm, multimode, SR, 3.3V, LC connector, with Digital Optical Monitoring | |||||||||
SFP-10G-LR | 10GBASE-LR SFP+ Module 1310 nm, single mode, LR, 3.3 V, with Digital Optical Monitoring | |||||||||
DS-SFP-FC8G-SW | 8 Gbit SFP+ Module 850 nm, multimode, SR, 3.3V, LC connector, with Digital Optical Monitoring | |||||||||
Footnote: | ||||||||||
[Approved Configurations] | - | |||||||||
[1] | Choose Up to Two Twinax Cables for Each Network Card Ordered
You may choose one or two twinax cables for each compatible PCIe network card ordered. The cables can be different lengths | |||||||||
[2] | Choose Up to Two SFP+ Modules for Each Network Card Ordered
You may choose one or two SFP+ optical modules cables for each compatible PCIe network card ordered. You would normally order two modules for connecting to the primary and redundant network switching equipment. With the SFP+ optical modules, you can use common fiber optic cables, widely available. |
GPU Cards (Optional)
Product ID (PID) | PID Description | Card Size | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
GPU PCIe Cards | ||||||||||
UCSC-GPU-VGXK1 | NVIDIA GRID K1 | Full-Height, double wide | - | |||||||
UCSC-GPU-VGXK2 | NVIDIA GRID K2 | Full-Height, double wide | - | |||||||
UCSC-GPU-K40 | NVIDIA Tesla K40 | Full-Height, double wide | - | |||||||
UCSC-GPU-M60 [1] | NVIDIA Tesla M60 | Full-Height, double wide | - | |||||||
UCSC-GPU-K80 [1] | NVIDIA K80 | Full-Height, double wide | - | |||||||
Available GPU Power Cables | ||||||||||
UCSC-GPUCBL-240M4 | C240 M4 GPU Power Cable | - | - | |||||||
Available GPU Power Cables (K80 and M60 only) | ||||||||||
UCS-300WKIT-240M4 | 300 Watt Cable, HS and Kit for UCS C240M4 Rack Server | - | - | |||||||
UCSC-C240-M4-LFF_GPU | ||||||||||
Footnote: | ||||||||||
[1] | You must order a kit with this GPU (UCS-300WKIT-240M4). It is a 300-Watt cable, HS and kit system for the UCS C240M4 rack server. Only one kit is needed for either one or two GPUs. | |||||||||
[2] | A power cable is not included with the M10. When ordering an M10, you must to purchase a power cable (UCSC-GPUCBL-240M4=) for the GPU. | |||||||||
[3] | CAUTION: Do not operate the C240 M4 server with the 300W GPU kit installed, but no GPU card installed. The kit has been designed to provide adequate airflow for cooling only when at least one GPU card is installed. | |||||||||
[4] | CAUTION: When using GPU cards, the operating temperature range is 32° to 95°F (0° to 35°C). | |||||||||
[5] | All GPU cards require two CPUs and a minimum of two power supplies in the server. 1400 W power supplies are recommended. Use the power calculator at the following link to determine the needed power based on the options chosen (CPUs, drives, memory, and so on): | |||||||||
[6] | Whenever you select a K1/K2/K40 or AMD GPU for this server, you must also select one power cable for each GPU selected. | |||||||||
[7] | Whenever you select a K80 GPU for this server, you must also select one power cable kit (for either one or two K80 GPUs). | |||||||||
[Caveats] | - | |||||||||
[-] | NVIDIA GPUs can support only less than 1 TB of total memory in the server. Do not install more than fourteen 64-GB DIMMs when using an NVIDIA GPU card in this server. Please note, this limitation does not apply to Pascal series GPU (P100) | |||||||||
[-] | NVIDIA GRID K1 and K2 GPUs can be mixed. No other GPU mixing is allowed. | |||||||||
[-] | If you order a K80 GPU, note the following:
| |||||||||
[-] | Slot 5 on riser card 2 is the required slot for the first GPU. | |||||||||
[-] | Slot 2 on riser card 1 is the secondary slot for a second GPU. The riser card 1 options that
are compatible with GPUs are:
| |||||||||
[-] | If you order one or two boot drives, you can order a maximum of one GPU and it must be installed in Riser 2 (UCSC-PCI-2-C240M4) slot 5. |
Available Power Supply
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
UCSC-PSU2V2-1400W | 1400W V2 AC Power Supply (200 - 240V) 2U & 4U C Series | |||||||||
UCSC-PSU2V2-1200W [1] | 1200W / 800W V2 AC Power Supply for 2U C-Series Servers | |||||||||
UCSC-PSU-930WDC | 930 W -48V DC Common Slot Power Supply for C-series servers | |||||||||
UCSC-PSU2V2-930DC | 930 W - 48V V2 DC Power Supply for 2U C-Series Servers | |||||||||
UCSC-PSU2V2-650W | 650W V2 AC Power Supply for C-Series Servers | |||||||||
Footnote: | ||||||||||
[1] | The C240 M4 server requires at least one power supply. A lightly loaded server may require one or two 650 W power supplies. A fully loaded server might need to be powered with two larger capacity power supplies. A server with one GPU requires at least two power supplies (1400 W power supplies are recommended). A server with two GPUs also requires at least two power supplies (1400 W power supplies are recommended | |||||||||
[2] | In a two power supply server, both power supplies must be identical |
Available Power Cords
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
R2XX-DMYMPWRCORD | No power cord (dummy PID to allow for a no power cord option) | |||||||||
CAB-N5K6A-NA | Power Cord, 200/240V 6A, North America | |||||||||
CAB-AC-L620-C13 | AC Power Cord, NEMA L6-20 - C13, 2M/6.5ft | |||||||||
CAB-C13-CBN | CABASY,WIRE,JUMPER CORD, 27"" L, C13/C14, 10A/250V | |||||||||
CAB-C13-C14-2M | CABASY,WIRE,JUMPER CORD, PWR, 2 Meter, C13/C14,10A/250V | |||||||||
CAB-C13-C14-AC | CORD,PWR,JMP,IEC60320/C14,IEC6 0320/C13, 3.0M | |||||||||
CAB-250V-10A-AR | Power Cord, 250V, 10A, Argentina 186571 2500 mm Cordset rating: 10 A, 250/500 V MAX Length: 8.2 ft Plug: EL 219 (IRAM 2073) Connector: EL 701 (IEC60320/C13) | |||||||||
CAB-9K10A-AU | Power Cord, 250VAC 10A 3112 Plug, Australia Plug: Cordset rating: 10 A, 250 V/500 V MAX Length: 2500mm 186580 Connector: EL 701C EL 210 (EN 60320/C15) (BS 1363A) 13 AMP fuse | |||||||||
CAB-250V-10A-CN | AC Power Cord - 250V, 10A - PRC | |||||||||
CAB-9K10A-EU | Power Cord, 250VAC 10A CEE 7/7 Plug, EU Connector: VSCC15 Cordset rating: 10A/16 A, 250 V Length: 8 ft 2 in. (2.5 m) Plug: M2511 186576 | |||||||||
CAB-250V-10A-ID | Power Cord, SFS, 250V, 10A, India | |||||||||
CAB-250V-10A-IS | Power Cord, SFS, 250V, 10A, Israel | |||||||||
CAB-9K10A-IT | Power Cord, 250VAC 10A CEI 23-16/VII Plug, Italy Plug: I/3G (CEI 23-16) Connector C15M (EN60320/C15 ) Cordset rating: 10 A, 250 V Length: 8 ft 2 in. (2.5 m) 186575 | |||||||||
CAB-9K10A-SW | Power Cord, 250VAC 10A MP232 Plug, Switzerland Plug: MP232-R Cordset rating: 10 A, 250 V Length: 8 ft. 2 in (2.5 m) 186578 Connector: IEC 60320 C15 | |||||||||
CAB-9K10A-UK | Power Cord, 250VAC 10A BS1363 Plug (13 A fuse), UK | |||||||||
CAB-9K12A-NA | Power Cord, 125VAC 13A NEMA 5-15 Plug, North America | |||||||||
CAB-250V-10A-BR | Power Cord - 250V, 10A - Brazil | |||||||||
CAB-48DC-40A-8AWG | C-Series -48VDC PSU Power Cord, 3.5M, 3 Wire, 8AWG, 40A | |||||||||
CAB-C19-C20-3M-JP | Power Cord C19-C20, 3M/10ft Japan PSE mark |
Available Cable Management Arms (CMA)
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
UCSC-CMAF-M4 | Reversible Cable Management Arm (CMA) for C220 M4 and M5 rack servers | |||||||||
Footnote: | ||||||||||
[Notes] | For more information about the tool-less rail kit and cable management arm, see the Cisco UCS C220 M5 Installation and Service Guide at this URL |
Available NIC Mode (Optional)
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
UCSC-DLOM-01 | Dedicated Mode BIOS setting for C-Series Servers | |||||||||
UCSC-CCARD-01 | Cisco Card Mode BIOS setting for C-Series Servers | |||||||||
Footnote: | ||||||||||
[1] | By default, the C220 M4 server NIC mode is configured to be Shared LOM Extended. This NIC mode allows any LOM port or adapter card port to be used to access the Cisco Integrated Management Controller (CIMC). The Cisco VIC card must be installed in a slot with NCSI support. | |||||||||
[2] | To change the default NIC mode to Dedicated, select the UCSC-DLOM-01 PID shown in Table Above. In Dedicated NIC mode, the CIMC can be accessed only through the dedicated management port. | |||||||||
[3] | To change the default NIC mode to Cisco Card Mode, select the UCSC-CCARD-01 PID shown in Table Above. In this mode, you can assign an IP address to the CIMC using DHCP and from there you can fully automate your deployment. |
Available Security Devices (Optional)
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
UCSX-TPM2-001 | Trusted Platform Module 1.2 SPI-based for UCS Servers | |||||||||
UCSX-TPM2-002 | Trusted Platform Module 2.0 for UCS servers | |||||||||
Footnote: | ||||||||||
[1] | A Trusted Platform Module (TPM) is a computer chip (microcontroller) that can securely store artifacts used to authenticate the platform (server). These artifacts can include passwords, certificates, or encryption keys. A TPM can also be used to store platform measurements that help ensure that the platform remains trustworthy. Authentication (ensuring that the platform can prove that it is what it claims to be) and attestation (a process helping to prove that a platform is trustworthy and has not been breached) are necessary steps to ensure safer computing in all environments. | |||||||||
[2] | A chassis intrusion switch gives a notification of any unauthorized mechanical access into the server. | |||||||||
[Notes] | The TPM module used in this system conforms to TPM v1.2 and 2.0, as defined by the Trusted Computing Group (TCG). It is also SPI-based. | |||||||||
[1] | TPM installation is supported after-factory. However, a TPM installs with a one-way screw and cannot be replaced, upgraded, or moved to another server. If a server with a TPM is returned, the replacement server must be ordered with a new TPM. |
Cisco SD Card Module (Optional)
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
UCS-SD-64G-S | 64 GB SD Card for UCS Servers | |||||||||
UCS-SD-32G-S | 32 GB SD Card for UCS Servers | |||||||||
Footnote: | ||||||||||
[Caveats] | - | |||||||||
[1] | Install either one or two SD cards | |||||||||
[2] | Do not mix SD card sizes |
Spare Parts List
Product ID (PID) | PID Description | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
UCSC-PCIF-01F= | PCIe Full Height blanking panel for UCS C-Series Server1 | |||||||||
UCSC-PCIF-C240M4= | C240 M4 PCIe Riser Blanking Panel1 | |||||||||
UCSC-PCI-2-C240M4= | C240 M4 PCIe Riser 2 Assembly1 | |||||||||
UCSC-PCI-1A-240M4= | C240 M4 PCIe Riser 1 Assembly (x8 slot + GPU)1 | |||||||||
UCSC-PCI-1B-240M4= | C240 M4 PCIe Riser 1 Assembly (3 x8 slots)1 | |||||||||
UCSC-PCI-1C-240M4= | C240 M4 PCIe Riser 1 Assembly (SATA Boot + 2 PCIe slots)1 | |||||||||
UCSC-IP-PCH-240M4= | Interposer board + cables for onboard PCH SATA 6G Embedded Software RAID1 | |||||||||
UCSC-MLOM-BLK= | MLOM Blanking Panel | |||||||||
UCS-240CBLMR12= | C240 M4 set of 2 cables for 12 HDD backplane chassis2 | |||||||||
UCSC-HS-C240M4= | Heat Sink for UCS C240 M4 Rack Server1 | |||||||||
UCS-CPU-LPCVR= | CPU load plate dust cover (for unpopulated CPU sockets) | |||||||||
UCS-GPUCBL-240M4= | C240 M4 GPU Power Cable (1 cable per GPU card)1 | |||||||||
N20-MBLIBATT= | Replacement Lithium Battery for Server Motherboard (CR2032)1 | |||||||||
UCSC-FAN-C240M4= | C240 M4 Fan Module (one) | |||||||||
UCSC-BAFF-C240M4= | C240 M4 Air Baffle Replacement Kit | |||||||||
UCSC-PSU-BLKP240= | Power Supply Blanking Panel for C240 M4 Servers1 | |||||||||
UCSC-RAILB-M4= | Tool-Less Ball Bearing Rail Kit for C220 M4 and C240 M4 rack servers | |||||||||
UCSC-CMAB-M4= | Reversible CMA for C240 M4 ball bearing rail kit | |||||||||
UCS-SD-32G-S= | 32 GB SD Card for UCS servers3 | |||||||||
UCS-SD-64G-S= | 64 GB SD Card for UCS servers2 | |||||||||
UCS-USBFLSHB-16GB= | 16GB Flash USB Drive | |||||||||
N20-BKVM= | KVM local IO cable for UCS servers console port | |||||||||
UCS-CPU-GREASE3= | M4 Server CPU thermal grease syringe - needed for heatsink seal4 | |||||||||
UCSX-HSCK= | UCS Processor Heat Sink Cleaning Kit (when replacing a CPU)3 | |||||||||
UCSC-MRAID-SC= | SuperCap for Cisco 12G SAS Modular RAID, including all cables. | |||||||||
UCS-300WKIT-240M4= | 300 Watt Cable and Kit for UCS C240M4 Rack Server Includes two heat sinks and | |||||||||
Footnote: | ||||||||||
[1] | This part is included/configured with your UCS server (in some cases, as determined by the configuration of | |||||||||
[2] | Required if ordering the RAID controller as a spare or to replace damaged cables | |||||||||
[3] | This SD card is blank. | |||||||||
[4] | This part should be ordered with the purchase of each optional or spare Intel Xeon E5-2600 v3 CPU processor kit |
PCIe Card Compatibility
PCIe Cards | Twinax Cables | Cisco SFP Modules | ||
---|---|---|---|---|
SFP-10G-SR | SFP-10G-LR | DS-SFP-FC8G-SW | ||
Converged Network Adapters (CNAs) | ||||
UCSC-PCIE-BSFP (Broadcom 57712 Dual Port 10Gb SFP+ w/TOE iSCSI) | Yes | Yes | No | |
UCSC-PCIE-CSC-02 (Cisco VIC 1225 Dual Port 10Gb SFP+ CNA) | Yes | Yes | Yes | No |
UCSC-PCIE-C10T-02 (Cisco VIC 1225T Dual Port 10GBaseT CNA) | No | No | No | No |
UCSC-MLOM-CSC-02 (Cisco UCS VIC1227 VIC MLOM - Dual Port 10Gb SFP+) | Yes | Yes | Yes | No |
UCSC-MLOM-C10T-02 (Cisco UCS VIC1227T VIC MLOM - Dual Port 10GBaseT) | No | No | No | No |
UCSC-PCIE-C40Q-02 (Cisco VIC 1285 Dual Port 40Gb QSFP CNA) | Yes | No [1] | No | |
UCSC-PCIE-C40Q-03 (Cisco VIC 1385 Dual Port 40Gb QSFP+ CNA w/RDMA) | Yes | No [1] | No | |
UCSC-PCIE-ESFP (Emulex OCe11102-FX dual-port 10 GbE FCoE CNA (Gen 3 CNA)) | Yes | Yes | No | |
UCSC-PCIE-QSFP (QLogic QLE8242-CU dual-port 10 GbE FCoE CNA) | Yes | Use Qlogic SFP | ||
UCSC-PCIE-B3SFP (Broadcom 57810 10Gb A-FEX SFP+ | Yes | Yes | No | |
UCSC-PCIE-Q8362 (Qlogic QLE8362 dual-port 10 GbE FCoE CNA) | Yes | Use Qlogic SFP | ||
UCSC-PCIE-E14102 (Emulex OCe14102-UX dual-port 10 GbE FCoE CNA) | Yes | Yes | No | |
Converged Network Adapters (CNAs) | ||||
N2XX-ABPCI01-M3 (Broadcom 5709 Dual-Port Ethernet PCIe Adapter for M3 Servers) | Use RJ45 Ethernet cable | |||
N2XX-ABPCI03-M3 (Broadcom 5709 Quad Port 10/100/1Gb NIC w/TOE iSCSI for M3 Servers | Use RJ45 Ethernet cable | |||
N2XX-AIPCI01 (Intel X520 Dual Port 10Gb SFP+ Adapter) | Yes | Use Intel SFP UCS-SFP-1WLR or UCS-SFP-1WSR | ||
UCSC-PCIE-ITG (Intel X540 Dual Port 10GBase-T Adapter) | Use RJ45 Ethernet cable | |||
UCSC-PCIE-IRJ45 (Intel i350 Quad Port 1Gb Adapter | Use RJ45 Ethernet cable | |||
UCSC-PCIE-IRJ45 (Intel i350 Quad Port 1Gb Adapter | Yes | No | No | |
Host Bus Adapters (HBAs) | ||||
N2XX-AEPCI03 (Emulex LPe 11002 Dual Port 4Gb Fibre Channel HBA | No | Preinstalled - do not change SFP | ||
N2XX-AEPCI05 (Emulex LPe 12002 Dual Port 8Gb Fibre Channel HBA) | No | Preinstalled - do not change SFP | ||
N2XX-AQPCI03 (QLogic QLE2462 Dual Port 4Gb Fibre Channel HBA) | No | Preinstalled - do not change SFP | ||
N2XX-AQPCI05 (QLogic QLE2562 Dual Port 8Gb Fibre Channel HBA) | No | Preinstalled - do not change SFP | ||
UCSC-PCIE-Q2672 (Qlogic QLE2672-CSC, 16Gb Fibre Channel HBA with SR Optics) | No | Preinstalled - do not change SFP | ||
UCSC-PCIE-E16002 (Emulex LPe16002-M6, 16Gb Fibre Channel HBA with SR Optics) | No | Preinstalled - do not change SFP | ||
Footnotes: | ||||
[1] | This card supports a 4x10 Gbps QSFP to SFP breakout fiber cable. |