| A | B | C | D | E | F | G | H | I | J | K | L | M | N | O | P | Q | R | S | T | U | V | W | X | Y | Z | |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 1 | Color Legend: | Identified Active Work Item | ||||||||||||||||||||||||
| 2 | Gaps / Possible Future Work | |||||||||||||||||||||||||
| 3 | Completed AI Items | Where is this driven at in OCP? | ||||||||||||||||||||||||
| 4 | Deprecated / No go / Not Applicable | |||||||||||||||||||||||||
| 5 | Project Sector | Project | Workitem | SC Priority | Current Status | Expected Completion | Narrative | PoC | New Workstream? | OCP Project | OCP Sub-Project | PL Priority | ||||||||||||||
| 6 | DC IT Infrastructure | Server | MHS Open System for AI | Early discussions on what areas / ingredients to focus on under the umbrella of MHS. | ​ | Taking similar approach as DC-MHs but expanding to include accelerators, semantic and scale-out NW, etc. | Shawn.Dube@dell.com brian.d.aspnes@intel.com | ​ | ||||||||||||||||||
| 7 | NIC 3.x - AI form factors | OCP NIC 3.x DSFF nearing completion. Need next steps. | ​ | Similar to smart NIC but AI accleration. | dchong@meta.com thomas1.ng@intel.com | ​ | ||||||||||||||||||||
| 8 | AI HW/SW Codesign | Whitepaper Pending | ​ | Whitepaper just published (late Sept 24). The follow through of graduating from FTI to OCP group to persue following up on the WP concepts is 4Q24,1Q25 | none | ​ | ||||||||||||||||||||
| 9 | Composable Infrastructure | CMS has launched an AI workstream. Need next steps. | ​ | manoj.wadekar@ocproject.net srajadnya@microsoft.com | Yes | ​ | ||||||||||||||||||||
| 10 | Quantization (OFP8) | 1.0 Complete - https://www.opencompute.org/documents/ocp-8-bit-floating-point-specification-ofp8-revision-1-0-2023-12-01-pdf-1 | ​ | ​ | ||||||||||||||||||||||
| 11 | Quantization (MX) | 1.0 Complete- https://www.opencompute.org/documents/ocp-microscaling-formats-mx-v1-0-spec-final-pdf | ​ | ​ | ||||||||||||||||||||||
| 12 | Chiplets for AI | Not Started. Need next steps. | ​ | AI specific chiplets and requirements such as faster BoW line rates, encouraging AI IP blocks to join, acclerator blocks etc... | Anu.Ramamurthy@microchip.com jawad@paloaltoelectron.com | ​ | ||||||||||||||||||||
| 13 | System in Package Security (SiP) | Caliptra....OCP and Linux Foundation. Could this be extended to support chiplets? Collaboration between Open Chiplets and Security project. | ​ | There is some concern about the ability about supply chain security such as counterfit components, misbehaving blocks and the leaving it easy to hack the SiP/components (substrate attacks etc...) An opportunity exists to define risks/best practices even define solutions in this area. | ​ | |||||||||||||||||||||
| 14 | Common/Standardized Hardware Abstraction layers | Not Started. Need next steps. | ​ | Software layers to abstract component(s) from AI technical stack such that developers may have hetrogeneous accleration components for AI without developing bespoke interfaces for each- this will grow the ecosystem. Could be in conjunction with AI SW-HW Co-design FTI group.Software layers and definitions. Work with OPF/HW MGt aspects such as firmware etc... | kasper.wszolek@intel.com | ​ | ||||||||||||||||||||
| 15 | In-Memory Computation | Not Started. Need next steps. | ​ | There is an FTS DCC group looking at this in additon to the CMS group. | manoj.wadekar@ocproject.net | ​ | ||||||||||||||||||||
| 16 | Rack Scale-Up / Scale-Out Network Protocol(s) | Protocols likely defined outside of OCP; ie: UALink, UEC, NVLink | ​ | ​ | ||||||||||||||||||||||
| 17 | Nvidia GB200 NVL-72 Rack | Not aligned with Open Systems for AI initiative. Was a fine contribution but not necessarily a point of long term innovation. | ​ | John Norton <johnn@nvidia.com> | ​ | |||||||||||||||||||||
| 18 | Common/Standardized Hardware Abstraction layers (MHS) | Combined with work above. | ​ | kasper.wszolek@intel.com | ​ | |||||||||||||||||||||
| 19 | OAI Re-design | Absorbed/accounted for as part of MHS Open System for AI thinking. | ​ | brian.d.aspnes@intel.com | ​ | |||||||||||||||||||||
| 20 | Cooling | ASIC Cooling Techniques | Not Started. Need next steps. | ​ | Probably splits into near (OCP) and long term (FTI) | tejas1.shah@intel.com | Yes | ​ | ||||||||||||||||||
| 21 | Networking | FTI Short Reach Optical Interconnect | Whitepaper in process (ETA: 5/15) | ​ | In progress with FTI group SROI | ron.swartzentruber@ocproject.net | ​ | |||||||||||||||||||
| 22 | CSIG (with UEC) | UEC is working on CSIC Spec (ETA: 10/24) | ​ | Starting up within Networking | barak.gafni@ocproject.net ravindra.sunkad@ocproject.net | Yes | ​ | |||||||||||||||||||
| 23 | Falcon + NVME + RDMA | Falcon 0.9 - Gathering Feedback - 1.0 Expected Q3 | ​ | Yes | ​ | |||||||||||||||||||||
| 24 | AI Driven Load balancing | Not Started. Need next steps. | ​ | Using telemetry/HW MGT to control networking and compute resources. Possibly requires Digital Twins. | none | Yes | ​ | |||||||||||||||||||
| 25 | Co-packaged Copper and/or Co-packaged Optics | Not Started. Need next steps. | ​ | One aspect (link types) In progress with FTI group Short Reach Optical Interconnect (SROI). Also could branch into useful implementations across server and networking | none | ​ | ||||||||||||||||||||
| 26 | 224G+ SerDes | This is an IEEE standards body piece of work. | ​ | There are changing requirements from 56G->112G->224G regarding board design requirements, layouts etc... This item is for study and any recommendations for implementation. Scope is to follow IEEE standards progress for for incorporation into next gen server design | thomas1.ng@intel.com | No | ​ | |||||||||||||||||||
| 27 | NIC 3.0 (AI NIC form factors) | Combine/grouped together with NIC 3.x - AI Form Factors above in Server project | ​ | thomas1.ng@intel.com | No | ​ | ||||||||||||||||||||
| 28 | Rack Scale-Up / Scale-Out Network(s) - Physical details | Combine/grouped together with NIC 3.x - AI Form Factors above in Server project | ​ | ​ | ||||||||||||||||||||||
| 29 | Storage | FTI Data Centric Computing | Ongoing whitepaper draft in FTI. New WG name for what used to be Computational Storage. | ​ | Minimizing data movement, intelligent clustering | yangseok.ki@samsung.com | ​ | |||||||||||||||||||
| 30 | Computational Storage | Folded into and renamed to Data Centric Computing covered above. | ​ | Multiple efforts have taken place over the years on this topic. While it has not been of immediate interest to OCP Community due to maturity, ambuigity etc... evolution of the idea continues and it's a potential architecture that we should take a defintive position on and if so, establish the community to implement | ​ | |||||||||||||||||||||
| 31 | Security | Chiplets (Chiplets as a Security Device) | Not Started. Need next steps. | ​ | There are layers to security and one layer may inclide application/data layer that can be accelerated in SiP/chiplet. Here is a distant example: https://ieeexplore.ieee.org/document/10286457 and https://ieeexplore.ieee.org/document/10466542 | bryankel@microsoft.com, jeffandersen@google.com | ​ | |||||||||||||||||||
| 32 | In-memory encryption | Not Started. Need next steps. | ​ | ​ | ||||||||||||||||||||||
| 33 | SAFE Outside Storage | Not Started. Need next steps. | ​ | bryankel@microsoft.com, jeffandersen@google.com | ​ | |||||||||||||||||||||
| 34 | Project Sector | Project | Workitem | SC Priority | Current Status | Expected Completion | Narrative | PoC | New Workstream? | OCP Project | OCP Sub-Project | PL Priority | ||||||||||||||
| 35 | DC Physical Infrastructure | DC Facilities | OCP Ready AI Version | Early Discussion | ​ | mailto:mark.dansie@opencompute.org | OCP Ready DC Recognition Program | ​ | ||||||||||||||||||
| 36 | DCF Sustainability (Green Concrete) | Started in Project | ​ | ​ | ||||||||||||||||||||||
| 37 | AI Driven Facility Management | Not Started | ​ | This might be best paired with Digital Twins. The goal is to automate and maximize efficiency via AI loops of control systems such as DCIM etc... | ​ | |||||||||||||||||||||
| 38 | Power Co-Generation (Alliances) | Not Started | ​ | Not enough power and/or not enough economically available power. Co-located generation will help with this | ​ | |||||||||||||||||||||
| 39 | Enhanced Heat Re-Use | Not Started | ​ | The first level hasn't been widely achieved- redirection of heat to coolants for downstream uses. More work to be done to make this widely adopted. Next level is doing more work (active work) with heat than transfering to a fluid (passive work) | jaime.comella@aq-compute.com | ​ | ||||||||||||||||||||
| 40 | TCS | ​ | Modularize the steps on Liquid Cooling such that instead of bespoke cooling designs for DCs, there is a far higher degree of standardization/commonality such that scales of economy help | Don.Mitchell@victaulic.com | ​ | |||||||||||||||||||||
| 41 | Rack and Power | AI Rack (200 - 250KW) (Power) | Scheduled for development after June 6th Update | ​ | harry.soin@ocproject.net, daniel.g.cartagena@intel.com | Yes | Rack and Power | HPR Rack for AI Infra | ​ | |||||||||||||||||
| 42 | AI Rack (200 - 250KW) (Cooling) | Not Started | ​ | gpcharest@meta.com | Yes | Rack and Power | HPR Rack for AI Infra | ​ | ||||||||||||||||||
| 43 | AI Rack (200 - 250KW) (Power Rail) | Not Started | ​ | gpcharest@meta.com | Yes | Rack and Power | HPR Rack for AI Infra | ​ | ||||||||||||||||||
| 44 | AI Rack (200 - 250KW) (Frame) | Scheduled for development after June 6th Update | ​ | steven.moore@ocproject.net | Yes | Rack and Power | HPR Rack for AI Infra | ​ | ||||||||||||||||||
| 45 | AI Rack 48VDC Power | Presented at OCP Global Summit 2024- Work in progress | ​ | ​ | ||||||||||||||||||||||
| 46 | Advanced Power Architecture AI Rack 400V | Presented at OCP Global Summit 2024- Work in progress | ​ | ​ | ||||||||||||||||||||||
| 47 | Advanced Power Architecture AI Rack 800V | Presented at OCP Global Summit 2024- Work in progress | ​ | ​ | ||||||||||||||||||||||
| 48 | AI Rack (300 - 500 KW) (Power) | Not Started | ​ | N+2 AI Acceleration components llikely require 2X the power/heat transfer than next gen (N) acceleration chips do. Racks need to be designed now to support this. | daniel.g.cartagena@intel.com | Yes | Rack and Power | HPR Rack for AI Infra | ​ | |||||||||||||||||
| 49 | AI Rack (300 - 500 KW) (Cooling) | Not Started | ​ | N+2 AI Acceleration components likely require 2X the power/heat transfer than next gen (N) acceleration chips do. Racks need to be designed now to support this. | Yes | Rack and Power | HPR Rack for AI Infra | ​ | ||||||||||||||||||
| 50 | AI Rack (300 - 500 KW) (Power Rail) | Not Started | ​ | Yes | Rack and Power | HPR Rack for AI Infra | ​ | |||||||||||||||||||
| 51 | AI Rack (300 - 500 KW) (Frame) | Not Started | ​ | Yes | Rack and Power | HPR Rack for AI Infra | ​ | |||||||||||||||||||
| 52 | AI Rack (300kW - 1MW) (Cooling) (OCP Global Summit MS- Mt.Diablo) | (OCP Global Summit MS- Mt.Diablo shared) | 2Q 2025 | ​ | ||||||||||||||||||||||
| 53 | AI Rack (300kW - 1MW) (Cooling) | (OCP Global Summit MS- Mt.Diablo shared) | 2Q 2025 | ​ | ||||||||||||||||||||||
| 54 | AI Rack (300kW - 1MW) (Cooling) | (OCP Global Summit MS- Mt.Diablo shared) | 2Q 2025 | ​ | ||||||||||||||||||||||
| 55 | Cooling | Advanced Cooling Facilities (500kw+) (400VDC) | Not Started | ​ | ​ | |||||||||||||||||||||
| 56 | Project Sector | Project | Workitem | SC Priority | Current Status | Expected Completion | Narrative | PoC | New Workstream? | OCP Project | OCP Sub-Project | PL Priority | ||||||||||||||
| 57 | Systems Managment | Hardware Management | RAS (Silent Data Corruption) | v1.0 contribution from non-public CLA workstream | ​ | jeyork@google.com | No | HW Management | Server Component Resilience WS | ​ | ||||||||||||||||
| 58 | RAS (GPU & Accelerator RAS Requirements) | v1.0 contribution from non-public CLA workstream | ​ | jeyork@google.com | No | HW Management | GPU Management WS | ​ | ||||||||||||||||||
| 59 | GPU HW Management (GPU & Accelerator Management Interfaces) | v1.0 contribution from non-public CLA workstream | ​ | jeyork@google.com | No | HW Management | GPU Management WS | ​ | ||||||||||||||||||
| 60 | GPU HW Management (Firmware Update Specification) | v1.0 contribution from non-public CLA workstream | ​ | jeyork@google.com | No | HW Management | GPU Management WS | ​ | ||||||||||||||||||
| 61 | Rack Management - Composable Disaggregated Infrastructure | Drafting OpenRMC v1.2 | ​ | Extending the prescription of rack management to include the composition service. The Redfish profile to v1.1 is posted. An GoLang implement of the rack manager which conforms to the profile is on the OCP respository. | hanwang@meta.com | No | HW Management | OpenRMC-DM Subproject | ​ | |||||||||||||||||
| 62 | Rack Management - PDU | Drafts exists, awaiting Rack & Power specification to reference | ​ | Prescribing management of the rack. Redfish profiles be drafted for the Rack PDU, and the Power Shelf | No | HW Management | Profiles WS | ​ | ||||||||||||||||||
| 63 | Liquid Cooling Management | Drafting Spec and Profile | ​ | Prescribing management of liquid cooling elements. Redfish profiles being drafted for Liquid Cooling Baseline, Cooling Distribution Unit (CDU), and the Rear Door Heat Exchanger | mike.jones@vertiv.com | No | Cooling Environments | HW Mgmt for Liquid Cooling WS | ​ | |||||||||||||||||
| 64 | Standardize Hardware Abstraction layers | [JL] Remove DC-MHS parenthetical. Organization doesn't need to be specified. Kasper states that this can be considered "under development". The MHS Plug and Play WS v0.3 has been released. | ​ | Software layers to abstract specific component from AI technical stack such that developers may have hetrogeneous accleration components for AI without developing bespoke interfaces for each (like system device drivers)- this will grow the ecosystem. With server group. Focus on the OPF/HW MGt aspects such as firmware etc. | kasper.wszolek@intel.com | No | Server | DC-MHS | ​ | |||||||||||||||||
| 65 | Fleet Management | Early drafting | ​ | Fleet management of 100K nodes. Covers lifecycle, configuration, telemetry, and control | sammy.nachimuthu@ocproject.net | No | HW Management | Cloud Service Model WS | ​ | |||||||||||||||||
| 66 | GPU HW Management Stack (Tooling & Telemetry) | Not Started or no visibility | ​ | Far more work can be done to create a hetrogeneous approach to AI accelerator management- meaning finer grained measurements, more meaningful measurements, and standardization across AI accelerator architectures/products | ​ | |||||||||||||||||||||
| 67 | AI Driven Automated Management (Digital Twins) | Not Started | ​ | Session to be held in AI track at Global Summit 2024 by Cadence team. Digital Twins to better manage the DC both for deployment and operations | ​ | |||||||||||||||||||||
| 68 | AI System Orchestration (Alliance with Linux Foundation) | Not Started | ​ | Integration of FOSS orchestration layers with OCP hardware. Some of the above steps would be towards this direction. | ​ | |||||||||||||||||||||
| 69 | Standarized Hardware Abstraction layers (OAI) | [JL] Combined with the DC-MHS entry (row 64) | ​ | Software layers for OAI to abstract component from AI technical stack such that developers may have hetrogeneous accleration components for AI without developing bespoke interfaces for each- this will grow the ecosystem. Could be in conjunction with AI SW-HW Co-design FTI group. Focus on the OPF/HW MGt aspects such as firmware etc... | ​ | |||||||||||||||||||||
| 70 | Test and Validation | AI Hardware System Benchmarking | Not Started | ​ | Need to standardize on measurements for AI performance. For example, how to define a 1EXFLOP system (what measures are used etc...) Definitions, templates and best practices | ​ | ||||||||||||||||||||
| 71 | ||||||||||||||||||||||||||
| 72 | ||||||||||||||||||||||||||
| 73 | HW MGT section has been verified by JL and he concurs with the work items | |||||||||||||||||||||||||
| 74 | ||||||||||||||||||||||||||
| 75 | ||||||||||||||||||||||||||
| 76 | ||||||||||||||||||||||||||
| 77 | ||||||||||||||||||||||||||
| 78 | ||||||||||||||||||||||||||
| 79 | ||||||||||||||||||||||||||
| 80 | ||||||||||||||||||||||||||
| 81 | ||||||||||||||||||||||||||
| 82 | ||||||||||||||||||||||||||
| 83 | ||||||||||||||||||||||||||
| 84 | ||||||||||||||||||||||||||
| 85 | ||||||||||||||||||||||||||
| 86 | ||||||||||||||||||||||||||
| 87 | ||||||||||||||||||||||||||
| 88 | ||||||||||||||||||||||||||
| 89 | ||||||||||||||||||||||||||
| 90 | ||||||||||||||||||||||||||
| 91 | ||||||||||||||||||||||||||
| 92 | ||||||||||||||||||||||||||
| 93 | ||||||||||||||||||||||||||
| 94 | ||||||||||||||||||||||||||
| 95 | ||||||||||||||||||||||||||
| 96 | ||||||||||||||||||||||||||
| 97 | ||||||||||||||||||||||||||
| 98 | ||||||||||||||||||||||||||
| 99 | ||||||||||||||||||||||||||
| 100 | ||||||||||||||||||||||||||