2805 Bowers Ave, Santa Clara, CA 95051 | 408-730-2275
sales@colfax-intl.com
My Colfax  

Colfax CX2265i-NVMe4-S8-XK8 2U Rackmount Server

 
 
  • 2x 3rd Gen Intel® Xeon® Scalable Processors
  • 2x NVIDIA Ampere Architecture GPUs
  • 32x DIMMS Support DDR4 RDIMM/LRDIMM
  • 8x 2.5" Gen4 U.2 NVMe Drive Bays

Hardware Features

  • 2x 3rd Gen Intel® Xeon® Scalable Processors
  • 2x NVIDIA Ampere Architecture GPUs
  • Intel C621A Chipset
  • 32x DIMMS Support DDR4 RDIMM/LRDIMM
  • 8x 2.5" Gen4 U.2 NVMe Drive Bays
  • Intel® Virtual RAID on CPU 7.5 (Intel® VROC 7.5) Support using one of the Three Types of VROC Keys
    - Standard Intel® VROC Key
    - Premium Intel® VROC Key
    - Intel® SSD Only VROC Key
  • 2x M.2 NVMe SSDs
  • Riser Slot #1 supports the following Riser Card options:
    - Two PCIe* slot riser card supporting (one) - FH/FL double-width slot (x16) + (one) - FH/HL single-width slot (x16)
  • Riser Slot #2 supports the following Riser Card options:
    - Two PCIe* slot riser card supporting (one) - FH/FL double-width slot (x16) + (one) - FH/HL single-width slot (x16)
  • Riser Slot #3 supports the following Riser Card options:
    - Two PCIe* slot riser card supporting (two) LP/HL single-width slots (x16)
  • 1x OCP Gen3 4.0 x16 Mezzanine Slot Supports Intel® Ethernet Network Adapter
  • Integrated Video Controller
  • Server Management:
    - Integrated Baseboard Management Controller (BMC)
    - Intelligent Platform Management Interface (IPMI) 2.0 Compliant
    - Redfish* Compliant
    - Support for Intel® Data Center Manager (DCM)
    - Support for Intel® Server Debug and Provisioning Tool (SDPTool)
    - Dedicated RJ45 1 GbE Management Port
    - Light Guided Diagnostics
  • 1x 2100W AC 80+ Titaniumum Efficiency Power Supply*
    * The system can have up to two power supply modules installed, supporting the following power configurations: 1+0, 1+1 redundant power, and 2+0 combined power
  • Optional Intel® Trusted Platform Module 2.0


Optional Features

Rack Mount Kit Options

  • Value Rack Mount Rail Kit (CYPHALFEXTRAIL):
    - 1U, 2U compatible
    - Tool-less chassis attachment
    - Tools required to attach rails to rack
    - Rack installation front and rear post distance adjustment from 660 mm to 838 mm
    - 560 mm travel distance
    - Half extension from rack
    - Support for front cover removal and fan replacement
    - 31 kg (68.34 lbs.) maximum support weight
    - No Cable Management Arm support
  • Premium Rail Kit with Cable Management Arm (CMA) Support (CYPFULLEXTRAIL):
    - 1U, 2U compatible
    - Tool-less installation
    - Rack installation front and rear post distance adjustment from 623 mm ~ 942 mm
    - 820 mm travel distance
    - Full extension from rack
    - 31 Kgs (68.34 lbs.) maximum supported weight
    - Support for Cable Management Arm AXXCMA2


PCIe Add-in Card Support

The server system supports a variety of riser card options for add-in card support as well as to enhance the base feature set of the system. These riser cards are available as accessory options for the server system. The system provides concurrent support for up to three PCIe riser cards. These riser cards provide concurrent support for up eight PCIe add-in cards.

PCI Express Bifurcation

The server system supports riser cards through riser slots identified as as Riser Slot #1, Riser Slot #2, and Riser Slot #3. The PCIe* data lanes for Riser Slot #1 are supported by CPU 0. The PCIe* data lanes for Riser Slot #2 and Riser Slot #3 are supported by CPU 1. A dual processor configuration is required when using Riser Slot #2 or Riser Slot #3.

The system supports the following PCIe bifurcation:

  • Add-in card slot 1 or slot 2 in 2-Slot PCIe* riser card (iPC – CYP2URISER1DBL) for Riser Slot #1:
    - x16/x8x8/x8x4x4/x4x4x8/x4x4x4x4
  • Add-in card slot 1 or slot 2 in 2-Slot PCIe* riser card (iPC – CYP2URISER2DBL) for Riser Slot #2:
    - x16/x8x8/x8x4x4/x4x4x8/x4x4x4x4


Two-Slot PCIe* Riser Card for Riser Slot #1 (iPC – CYP2URISER1DBL)

The two-slot PCIe* riser card option supports:

  • One FH/FL double-width slot (x16 electrical, x16 mechanical)
  • One FH/HL single-width slot (x16 electrical, x16 mechanical)

Two-Slot PCIe* Riser Card for Riser Slot #2 (iPC – CYP2URISER2DBL)

The two-slot PCIe* riser card option supports:

  • One FH/FL double-width slot (x16 electrical, x16 mechanical)
  • One FH/HL single-width slot (x16 electrical, x16 mechanical)

Two-Slot PCIe* Riser Card for Riser Slot #3 (iPC – CYP2URISER3STD)

The two-slot PCIe* riser card option supports:

  • Two LP/HL single-width slots (x16 mechanical, x8 electrical)


Intel® Virtual RAID on Chip (Intel® VROC) for NVMe

Intel® VROC supports the following:

  • I/O processor with controller (ROC) and DRAM
  • Protected write-back cache – software and hardware that allows recovery from a double fault
  • Isolated storage devices from operating system for error handling
  • Protected R5 data from operating system crash
  • NVMe SSD hot plug and surprise removal on processor PCIe lanes
  • LED management for PCIe attached storage
  • RAID/storage management using Representational State Transfer (RESTful) application programming interfaces (APIs)
  • Graphical user interface (GUI) for Linux
  • 4K native NVMe SSD support


VROC Upgrade Key - Supported NVMe RAID Features



Intel® Volume Management Device (Intel® VMD) 2.0 for NVMe

Intel® Volume Management Device (Intel® VMD) is hardware logic inside the processor root complex to help manage PCIe NVMe SSDs. It provides robust hot plug support and status LED management. This allows servicing of storage system NVMe SSD media without fear of system crashes or hangs when ejecting or inserting NVMe SSD devices on the PCIe bus.

Intel® VMD handles the physical management of NVMe storage devices as a stand-alone function but can be enhanced when Intel® Virtual RAID on CPU (Intel® VROC) support options are enabled to implement RAID-based storage systems.

  • Hardware is integrated inside the processor PCIe root complex
  • Entire PCIe trees are mapped into their own address spaces (domains)
  • Each domain manages x16 PCIe lanes
  • Can be enabled/disabled in BIOS Setup at x4 lane granularity
  • Driver sets up/manages the domain (enumerate, event/error handling)
  • May load an additional child device driver that is Intel® VMD aware
  • Hot plug support - hot insert array of PCIe NVMe SSDs
  • Support for PCIe NVMe SSDs only (no network interface controllers (NICs), graphics cards, and so on)
  • Maximum of 128 PCIe bus numbers per domain
  • Support for Management Component Transport Protocol (MCTP) over SMBus only
  • Support for MMIO only (no port-mapped I/O)
  • Does not support NTB, Quick Data Tech, Intel® Omni-Path Architecture (Intel® OPA), or SR-IOV
  • Correctable errors do not bring down the system
  • Intel® VMD only manages devices on PCIe lanes routed directly from the processor or PCH chipset
  • When Intel® VMD is enabled, the BIOS does not enumerate devices that are behind Intel® VMD. The Intel® VMD-enabled driver is responsible for enumerating these devices and exposing them to the host


Intel® Trusted Platform Module (TPM) 2.0

A TPM is a hardware-based security device that addresses the growing concern on boot process integrity and offers better data protection. TPM protects the system start-up process by ensuring it is tamper-free before releasing system control to the operating system. A TPM device provides secured storage to store data, such as security keys and passwords. In addition, a TPM device has encryption and hash functions.

AXXTPMENC8 implements TPM as per TPM PC Client specifications revision 2.0 by the Trusted Computing Group (TCG)



Intel, the Intel logo, Xeon, and Xeon Inside are trademarks or registered trademarks of Intel Corporation in the U.S. and other countries.

Server Management

Overview

The server uses the baseboard management controller (BMC) features of an ASpeed* AST2500 server management processor. The BMC supports multiple system management features including intra-system sensor monitoring, fan speed control, system power management, and system error handling and messaging. It also provides remote platform management capabilities including remote access, monitoring, logging, and alerting features.

In support of system management, the system includes a dedicated management port and support for two system management tiers and optional system management software.

  • Standard management features (Included)
  • Advanced management features (Optional)
  • Intel® Data Center Manager (DCM) support (Optional)

Remote Management Port
The server board includes a dedicated 1 Gb/s RJ45 management port used to access embedded system management features remotely.

Standard System Management Features
The following system management features are supported by default.

  • Virtual KVM over HTML5
  • Integrated BMC Web Console
  • Redfish
  • IPMI 2.0
    • Node Manager
  • Out-of-band BIOS/BMC Update and Configuration
  • System Inventory
  • Autonomous Debug Log

Advanced Management Features
Advanced manageability features are supported over all NIC ports enabled for server manageability. This includes baseboard integrated BMC-shared NICs, which share network bandwidth with the host system, as well as the LAN channel provided by the onboard Intel® Dedicated Server Management NIC.

  • Software Key to enable features
  • Included single system license for Intel® Data Center Manager (Intel® DCM)
    • Intel® Data Center Manager (Intel® DCM) is a software solution that collects and analyzes the real-time health, power, and thermals of a variety of devices in data centers helping you improve the efficiency and uptime.
  • Virtual Media Image Redirection (HTML5 and Java*)
  • Virtual Media over network share and local folder
  • Active Directory support
  • Full system firmware update including drives, memory, and RAID (Tentative Availability Q4 2021)
  • Storage and network device monitoring (Tentative Availability Q4 2021)
  • Out-of-band hardware RAID Management for latest Intel® RAID cards (Tentative Availability Q4 2021)

More Information
Download Integrated BMC Web Console User Guide


Intel® Data Center Manager (Intel® DCM)

Intel® DCM is a solution for out-of-band monitoring and managing the health, power, and thermals of servers and a variety of other types of devices.

What can you do with Intel® DCM?

  • Automate health monitoring
  • Improve system manageability
  • Simplify capacity planning
  • Identify underutilized servers
  • Measure energy use by device
  • Pinpoint power/thermal issues
  • Create power-aware job scheduling tasks
  • Increase rack densities
  • Set power policies and caps
  • Improve data center thermal profile
  • Optimize application power consumption
  • Avoid expensive PDUs and smart power strips

More Information
Download Intel® Data Center Manager Product Brief
Download Intel® Data Center Manager Console User Guide



Intel® DCM Use Cases

Rack Provisioning

Find new ways to increase rack density.

Intelligent Power

Collect real time data without deploying costly redundant infrastructure by replacing intelligent power distribution units.

Disaster Avoidance

With real-time monitoring and management, it's possible to reduce power failures and other disasters.

Equipment Scheduling

Increase your ability to meet workload demands with equipment scheduling, and make your data center do more.

Build Real-Time Thermal Maps

Build real-time thermal maps to avoid the guesswork that leads to undercooling or overcooling.

Ghost Servers

Identify ghost servers, and get data center power usage under control.

Intel® Power Thermal Aware Solution

Identify energy efficiency issues in the data center to avoid service delays and gain savings.

Granular Rack-Level Thermal Monitoring

Enable the Intel® DCM to recognize an out-of-range temperature reading and allow the user to take immediate action.

Granular Server-Level Thermal Monitoring

Get greater granular server-level thermal visibility, so when temperatures rise, it registers with the Intel® DCM.

Predictive Detection of Cooling Anomalies

Predict cooling issues before they happen with a patented algorithm that detects anomalies in time to be resolved before a thermal issue occurs.

Server health Management

Enable server health management with real-time sub-component monitoring, error detection, proactive health management, and server firmware synchronization.

Updating Firmware of Intel® Data Center Blocks

Monitor and update the firmware of data center systems with Intel® DCM.

Intel® Memory Failure Prediction

Through multi-dimensional model and algorithms, DIMM errors are mined at the micro-level to assign health scores and identify future failures in real time.

Technical Specifications

Dimensions (HxWxL) • 2U Rackmount
• 3.42" x 17.56" x 30.3"
• 87mm x 446mm x 770mm
CPU • Dual Socket-P4 LGA4189
• Support for 3rd Gen Intel® Xeon® Scalable Processors
• Max TDP up to 205 W
• UPI links: up to three at 11.2 GT/s (Platinum and Gold families) or up to two at 10.4 GT/s (Silver family)

Note: Supported 3rd Gen Intel® Xeon® Scalable processor SKUs must Not end in (H), (L), (U), or (Q). All other processor SKUs are supported.
Chipset • Intel® C621A
Memory • 32x DIMM slots
- 16 DIMM slots per processor, eight memory channels per processor
- Two DIMMs per channel
• Supports Registered DDR4 (RDIMM), 3DS-RDIMM, Load Reduced DDR4 (LRDIMM), 3DS-LRDIMM
• Intel® Optane™ persistent memory 200 series
• Memory capacity
- Up to 6 TB per processor (processor SKU dependent)
• Memory data transfer rates
- Up to 3200 MT/s at one or two DIMMs per channel (processor SKU dependent)
• DDR4 standard voltage of 1.2V
Riser Support Concurrent support for up to three riser cards with support for up to eight PCIe* add-in cards. In the below description FH = Full Height, FL = Full Length, HL =Half Length, LP = Low Profile

Riser Slot #1:
• Riser Slot #1 supports x32 PCIe* lanes, routed from CPU 0
• PCIe 4.0 support for up to 64 GB/s

Riser Slot #1 supports the following Intel Riser Card option:
• Three PCIe* slot riser card supporting (one) - FH/FL single-width slot (x16 electrical, x16 mechanical) + (one) - FH/FL single-width slot (x8 electrical, x16 mechanical) + (one) - FH/HL single-width slot (x8 electrical, x8 mechanical) iPC – CYP2URISER1STD
• Two PCIe* slot riser card supporting (one) - FH/FL double-width slot (x16 electrical, x16 mechanical) + (one) - FH/HL single-width slot (x16 electrical, x16 mechanical) iPC – CYP2URISER1DBL
• NVMe* riser card supporting (one) – HL or FL single-width slot (x16 electrical, x16 mechanical) + (two) - x8 PCIe* NVMe* SlimSAS* connectors, each with a re-timer. iPC – CYP2URISER1RTM

Riser Slot #2:
• Riser Slot #2 supports X32 PCIe lanes routed from CPU 1
• PCIe 4.0 support for up to 64 GB/s

Riser Slot #2 supports the following Intel Riser Card option:
• Three PCIe* slot riser card supporting (one) - FH/FL single-width slot (x16 electrical, x16 mechanical) + (one) - FH/FL single-width slot (x8 electrical, x16 mechanical) + (one) FH/HL single-width slot (x8 electrical, x8 mechanical) iPC – CYP2URISER2STD
• Two PCIe* slot riser card supporting (one) - FH/FL double-width slot (x16 electrical, x16 mechanical) + (one) - FH/HL single-width slot (x16 electrical, x16 mechanical) iPC – CYP2URISER2DBL

Riser Slot #3:
• Riser Slot #3 supports x16 PCIe* lanes routed from CPU 1
• PCIe 4.0 support for up to 32 GB/s

Riser Slot #3 supports the following Intel Riser Card option:
• Two PCIe* slot riser card supporting (two) LP/HL single-width slots (x16 mechanical, x8 electrical) iPC – CYP2URISER3STD
• NVMe riser card supporting (two) – PCIe NVMe SlimSAS connectors iPC – CYPRISER3RTM
Open Compute Project (OCP) Adapter Support Onboard x16 PCIe 4.0 OCP 3.0 Mezzanine connector (Small Form-Factor) supports the following Intel accessory options:
• Dual port, RJ45, 10/1 GbE - iPC- X710T2LOCPV3
• Quad port, SFP+ DA, 4x 10 GbE - iPC- X710DA4OCPV3
• Dual Port, QSFP28 100/50/25/10 GbE - iPC- E810CQDA2OCPV3
• Dual Port, SFP28 25/10 GbE - iPC-E810XXVDA2OCPV3
PCIe NVMe Support • Support for up to 10 PCIe NVMe Interconnects
- Eight server board SlimSAS connectors, four per processor
- Two M.2 NVMe/SATA connectors
• Additional NVMe support through select Riser Card options (See Riser Card Support)
• Intel® Volume Management Device 2.0 (Intel® VMD 2.0) support
• Intel® Virtual RAID on CPU 7.5 (Intel® VROC 7.5) support using one of the three types of VROC keys (available as an Intel accessory option)
SATA • 10 x SATA III ports (6 Gb/s, 3 Gb/s and 1.5 Gb/s transfer rates supported)
- Two M.2 connectors – SATA / PCIe
- Two 4-port Mini-SAS HD (SFF-8643) connectors
USB • Three USB 3.0 connectors on the back panel
• One USB 3.0 and one USB 2.0 connector on the front panel
• One USB 2.0 internal Type-A connector
Serial • One external RJ-45 Serial Port A connector on the back panel
• One internal DH-10 Serial Port B header for optional front or rear serial port support. The port follows the DTK pinout specifications
Video • Integrated 2D video controller
• 128MB of DDR4 video memory
• One VGA DB-15 external connector in the back
Server Management • Integrated Baseboard Management Controller
• Intelligent Platform Management Interface (IPMI) 2.0 Compliant
• Redfish* Compliant
• Support for Intel® Data Center Manager (DCM)
• Support for Intel® Server Debug and Provisioning Tool (SDPTool)
• Dedicated RJ45 1 GbE Management Port
• Light Guided Diagnostics
Security Support • Intel® Platform Firmware Resilience (Intel® PFR) technology with an I2C interface
• Intel® Software Guard Extensions (Intel® SGX) • Intel® CBnT – Converged Intel® Boot Guard and Trusted Execution Technology (Intel® TXT) • Intel® Total Memory Encryption (Intel® TME) • Trusted platform module 2.0 – iPC J33567-151 (accessory option)
Storage Bay • 8 x 2.5" SAS/SATA/NVMe* hot swap drive bays
Power Supply The server system can have up to two power supply modules installed, supporting the following power configurations: 1+0, 1+1 redundant power, and 2+0 combined power

• 1 x 1300W / 1600W / 2100W AC power supply
• 80 Plus Titanium
System Fans • Six managed 60 mm hot swap capable system fans
• Integrated fans included with each installed power supply module
BIOS • Unified Extensible Firmware Interface (UEFI)-based BIOS (legacy boot not supported


Block Diagram


Images


$4,763.18
 

  Qty.Colfax CX2265i-NVMe4-S8-XK8 2U Rackmount Server, Cost As Configured $4,763.18
Base Platform 1
Front Bezel 1
Management 1
Power Supply 1    1

Set next    items (  ) like this
Power Supply 2    1
Rackmount Kit 1
Cable Management 1
Primary CPU 1
Secondary CPU 1
CPU 1 - Socket 1 of 16    1

Set next    items (  ) like this
CPU 1 - Socket 2 of 16    1
CPU 1 - Socket 3 of 16    1
CPU 1 - Socket 4 of 16    1
CPU 1 - Socket 5 of 16    1
CPU 1 - Socket 6 of 16    1
CPU 1 - Socket 7 of 16    1
CPU 1 - Socket 8 of 16    1
CPU 1 - Socket 9 of 16    1
CPU 1 - Socket 10 of 16    1
CPU 1 - Socket 11 of 16    1
CPU 1 - Socket 12 of 16    1
CPU 1 - Socket 13 of 16    1
CPU 1 - Socket 14 of 16    1
CPU 1 - Socket 15 of 16    1
CPU 1 - Socket 16 of 16    1
CPU 2 - Socket 1 of 16    1

Set next    items (  ) like this
CPU 2 - Socket 2 of 16    1
CPU 2 - Socket 3 of 16    1
CPU 2 - Socket 4 of 16    1
CPU 2 - Socket 5 of 16    1
CPU 2 - Socket 6 of 16    1
CPU 2 - Socket 7 of 16    1
CPU 2 - Socket 8 of 16    1
CPU 2 - Socket 9 of 16    1
CPU 2 - Socket 10 of 16    1
CPU 2 - Socket 11 of 16    1
CPU 2 - Socket 12 of 16    1
CPU 2 - Socket 13 of 16    1
CPU 2 - Socket 14 of 16    1
CPU 2 - Socket 15 of 16    1
CPU 2 - Socket 16 of 16    1
M.2 Drive 1    1

Set next    items (  ) like this
M.2 Drive 2    1
NVM-E Drive 1    1

Set next    items (  ) like this
NVM-E Drive 2    1
NVM-E Drive 3    1
NVM-E Drive 4    1
NVM-E Drive 5    1
NVM-E Drive 6    1
NVM-E Drive 7    1
NVM-E Drive 8    1
NVM-E Cable Kit (1-8) 1
Intel VROC 1
OCP 3.0 Networking 1
Riser Card 1 1
Riser Card 2 1
Riser Card 3 1
Nvidia GPU 1    1

Set next    items (  ) like this
Nvidia GPU 2    1
Infiniband HBA 1
Ethernet HBA 1
TPM Module 1
Operating System SW 1
   Colfax CX2265i-NVMe4-S8-XK8 2U Rackmount Server, Cost As Configured $4,763.18