By: Daniel B (nothanks.delete@this.nothanks.com), July 23, 2022 4:15 pm
Room: Moderated Discussions
Ben T (ben.delete@this.noemail.com) on July 22, 2022 10:02 pm wrote:
> > Do people here think the ideal public cloud infrastructure would be
> > a cluster in a box rather than a single symmetric multiprocessor
> > running virtual machines from multiple users? For example, imagine
> > a 1U chassis containing 12 processor chips and 12 hot-swap
> > 2.5” U.2 NVMe drives. Each processor chip could include:
> >
> > 8 performance CPU cores plus caches
> > interfaces to two 128-bit LPDDR DRAMs
> > 4 PCIe lanes to connect to one NVMe drive
> > a 10G ethernet connection
> >
> > If the minimum unit a cloud user could get is 8 cores, they would not
> > be sharing this chip with anyone else so the risk of data leakage
> > between users would be greatly reduced. If it is possible to combine
> > these processor units for bigger single tasks but still keep them
> > partitionable into independent units for small tasks, that would be even better.
>
> Here is one way to put 12 processor packages in a 1U server. To fit in a standard 19” rack, the width of
> the motherboard has to be at most 17” (43.18cm). To have a single row of processor packages across the
> width of the chassis, each processor package could be about 3.4cm wide with a .2cm gap between packages.
> A processor package containing two LPDDR5X DRAM packages would be rectangular with the processor die in the
> middle and an LPDDR5X DRAM package on two opposite sides of the processor die. An LPDDR5X DRAM package is
> 1.4cm x 1.4cm. Apple’s M1 Max die is 2.16cm x 1.99cm so this all fits comfortably in a 3.4cm wide package.
> The long side of the rectangular processor package would be parallel to the air flow in the chassis. An alternative
> would be to have two rows of six processor packages (P) with a big gap between each package in a row so that
> cool air can reach the row of packages farthest from the air intake, like this:
>
> P_P_P_P_P_P
> _P_P_P_P_P_P
>
> A reason to have 12 processor packages is that the number of 2.5” drives that fit across the front
> of a chassis is a multiple of 12 (12 in a 1U chassis, 24 in a 2U chassis). Less fan power is needed
> when the heat sources are spread out instead of concentrated in one or two processor sockets.
>
> Does this seem practical?
Looks like you re-invented the micro-server approach, which has already failed? Why not just use cartridges of micro/pico/nano whatever arbitrary name server cards share an electrical and IO subsystems?
Couple of things:
- loss of instance versatility as this infrastructure will not be good at serving both low-end and high-end needs
- loss of vcpu performance as smaller thermal envelope and smaller shared uncore and memory
- added complexity in supply chain, technical support and inventory management
- added complexity in operations as you'll need new system images, update to orchestration tools
- because of smaller resource pools, more difficult in achieving high utilization to drive economics and energy efficiency
- you seem to be thinking about thermal management as a driver - once that becomes a gating factor, it's likely liquid cooling will be the response rather than generating more IT hardware swell
For some applications, what you describe may make sense at scale, such as web frontends (remember, Facebook made Intel create uniprocessor Xeon Ds). For general purpose cloud infrastructure, it does not seem like an attractive direction.
Bottom line is, the direction is ever larger resource pools, and cloud wants ever larger resource pools at the most economic point.
> > Do people here think the ideal public cloud infrastructure would be
> > a cluster in a box rather than a single symmetric multiprocessor
> > running virtual machines from multiple users? For example, imagine
> > a 1U chassis containing 12 processor chips and 12 hot-swap
> > 2.5” U.2 NVMe drives. Each processor chip could include:
> >
> > 8 performance CPU cores plus caches
> > interfaces to two 128-bit LPDDR DRAMs
> > 4 PCIe lanes to connect to one NVMe drive
> > a 10G ethernet connection
> >
> > If the minimum unit a cloud user could get is 8 cores, they would not
> > be sharing this chip with anyone else so the risk of data leakage
> > between users would be greatly reduced. If it is possible to combine
> > these processor units for bigger single tasks but still keep them
> > partitionable into independent units for small tasks, that would be even better.
>
> Here is one way to put 12 processor packages in a 1U server. To fit in a standard 19” rack, the width of
> the motherboard has to be at most 17” (43.18cm). To have a single row of processor packages across the
> width of the chassis, each processor package could be about 3.4cm wide with a .2cm gap between packages.
> A processor package containing two LPDDR5X DRAM packages would be rectangular with the processor die in the
> middle and an LPDDR5X DRAM package on two opposite sides of the processor die. An LPDDR5X DRAM package is
> 1.4cm x 1.4cm. Apple’s M1 Max die is 2.16cm x 1.99cm so this all fits comfortably in a 3.4cm wide package.
> The long side of the rectangular processor package would be parallel to the air flow in the chassis. An alternative
> would be to have two rows of six processor packages (P) with a big gap between each package in a row so that
> cool air can reach the row of packages farthest from the air intake, like this:
>
> P_P_P_P_P_P
> _P_P_P_P_P_P
>
> A reason to have 12 processor packages is that the number of 2.5” drives that fit across the front
> of a chassis is a multiple of 12 (12 in a 1U chassis, 24 in a 2U chassis). Less fan power is needed
> when the heat sources are spread out instead of concentrated in one or two processor sockets.
>
> Does this seem practical?
Looks like you re-invented the micro-server approach, which has already failed? Why not just use cartridges of micro/pico/nano whatever arbitrary name server cards share an electrical and IO subsystems?
Couple of things:
- loss of instance versatility as this infrastructure will not be good at serving both low-end and high-end needs
- loss of vcpu performance as smaller thermal envelope and smaller shared uncore and memory
- added complexity in supply chain, technical support and inventory management
- added complexity in operations as you'll need new system images, update to orchestration tools
- because of smaller resource pools, more difficult in achieving high utilization to drive economics and energy efficiency
- you seem to be thinking about thermal management as a driver - once that becomes a gating factor, it's likely liquid cooling will be the response rather than generating more IT hardware swell
For some applications, what you describe may make sense at scale, such as web frontends (remember, Facebook made Intel create uniprocessor Xeon Ds). For general purpose cloud infrastructure, it does not seem like an attractive direction.
Bottom line is, the direction is ever larger resource pools, and cloud wants ever larger resource pools at the most economic point.
Topic | Posted By | Date |
---|---|---|
Retbleed | anonymous2 | 2022/07/13 03:14 PM |
Retbleed | anon2 | 2022/07/13 10:03 PM |
Retbleed | Adrian | 2022/07/14 12:05 AM |
Retbleed | Anon4 | 2022/07/14 02:17 PM |
Retbleed | anon2 | 2022/07/14 04:29 PM |
Retbleed | Anon4 | 2022/07/14 05:05 PM |
Retbleed | anon2 | 2022/07/14 05:37 PM |
Retbleed | anon2 | 2022/07/14 06:40 PM |
Retbleed | dmcq | 2022/07/15 04:54 AM |
Retbleed | anon2 | 2022/07/17 07:17 AM |
Retbleed | Michael S | 2022/07/15 07:08 AM |
Retbleed | Ben T | 2022/07/16 05:06 AM |
Retbleed | Michael S | 2022/07/16 11:41 AM |
Public cloud infrastructure | Ben T | 2022/07/16 04:50 PM |
Public cloud infrastructure | Rayla | 2022/07/16 09:15 PM |
Public cloud infrastructure | me | 2022/07/17 09:19 AM |
Public cloud infrastructure | Brett | 2022/07/18 12:38 PM |
Public cloud infrastructure | Adrian | 2022/07/18 01:19 PM |
Public cloud infrastructure | me | 2022/07/18 03:54 PM |
Public cloud infrastructure | Brett | 2022/07/20 03:35 PM |
Public cloud infrastructure | Brett | 2022/07/21 01:18 PM |
Public cloud infrastructure | inthestratosphere | 2022/07/21 02:46 PM |
Public cloud infrastructure | Brett | 2022/07/21 10:38 PM |
What’s needed for a viable Apple server? | Ben T | 2022/07/22 05:31 AM |
What’s needed for a viable Apple server? | Michael S | 2022/07/22 09:09 AM |
More DRAM capacity? | Mark Roulo | 2022/07/22 09:48 AM |
More DRAM capacity? | Doug S | 2022/07/22 11:05 AM |
More DRAM capacity? | Mark Roulo | 2022/07/22 11:20 AM |
More DRAM capacity? | Doug S | 2022/07/22 01:48 PM |
More DRAM capacity? | Wes Felter | 2022/07/22 04:49 PM |
Public cloud infrastructure | anon2 | 2022/07/18 04:25 PM |
Putting 12 processor packages in a 1U server | Ben T | 2022/07/22 10:02 PM |
Putting 12 processor packages in a 1U server | rwessel | 2022/07/23 07:15 AM |
Putting 12 processor packages in a 1U server | Daniel B | 2022/07/23 04:15 PM |
Putting 12 processor packages in a 1U server | Ben T | 2022/07/24 05:29 AM |
Multi-system cluster design space | Paul A. Clayton | 2022/07/24 08:49 AM |
Retbleed | Anon4 | 2022/07/15 03:00 AM |
Retbleed | Michael S | 2022/07/15 06:59 AM |
Retbleed | --- | 2022/07/15 11:14 AM |