By: gallier2 (gallier2.delete@this.gmx.de), March 5, 2021 3:22 am
Room: Moderated Discussions
Linus Torvalds (torvalds.delete@this.linux-foundation.org) on March 4, 2021 8:21 pm wrote:
> David Hess (davidwhess.delete@this.gmail.com) on March 4, 2021 1:44 pm wrote:
> >
> > I have not noticed any delay on my workstation which has 64G.
>
> The people who were hurting are the people with many hundreds of gigs (or terabytes)
> of RAM, and have five-nines availability guarantees, and want their scheduled
> maintenance reboots to take on the order of seconds, not minutes.
>
> Yes, yes, they obviously actually use many other strategies (ie multiple nodes with fail-over etc) because
> generally there's no way they'll hit the reboot speed requirements even for scheduled downtime, much less
> for when something actually goes wrong. But they do tend to want "fail fast, and get back up fast".
>
> People like that absolutely hated the traditional SCSI disk spin-up times, and they found fsck times
> completely unacceptable, and they want the RAM to be scrubbed and initialized incrementally so that you
> can boot and get the system up and the loads started ASAP. Keep scrubbing and initializing more RAM in
> parallel, because it's going to take a while before you're under full load and need it all anyway.
>
> Those people tend to be a fairly special breed.
>
> There's very much the other kind of server person, who is used to (and so resigned as to expect) the server
> reboot taking possibly tens of minutes, because that's just what they always did - spinning up one rotating
> platter of rust at a time to avoid the power surge of spinning them all up at the same time etc etc.
>
> I'm not even kidding. That whole "spin up one disk at a time" is
> not some exaggeration. It was standard practice in many places.
>
> Of course, it's largely historical, and had real roots - those disks used to be massive, and spinning
> them up really was a big deal. It's just that the practice of spinning disks up one at a time actually
> went on in many places long after those physically massive disk systems were long long gone.
>
> So you have this odd dichotomy of boot speed expectations in the big iron world. Some want it
> to be really instant and quick. And some are perfectly happy with ridiculously slow boots.
>
> I've had access to big powerful hardware that was almost entirely useless for kernel development, because
> while they could compile a new kernel really quickly, they then took literally several minutes to boot
> the damn thing. So a much slower machine was actually much better at doing any actual testing on.
>
Yeah, I had the same issue with a HP-UX machine (HP-747i) in the 90s for developing for a 68030 based VME communication card. Compiling and linking the kernel driver took seconds, rebooting without fsck took 15 minutes, with fsck 25 and it was really easy to crash the machine when communicating over the VME bus. What a luck that I could do most of the 030 part of the code on my Atari TT, otherwise I would probably not have survived that project.
> David Hess (davidwhess.delete@this.gmail.com) on March 4, 2021 1:44 pm wrote:
> >
> > I have not noticed any delay on my workstation which has 64G.
>
> The people who were hurting are the people with many hundreds of gigs (or terabytes)
> of RAM, and have five-nines availability guarantees, and want their scheduled
> maintenance reboots to take on the order of seconds, not minutes.
>
> Yes, yes, they obviously actually use many other strategies (ie multiple nodes with fail-over etc) because
> generally there's no way they'll hit the reboot speed requirements even for scheduled downtime, much less
> for when something actually goes wrong. But they do tend to want "fail fast, and get back up fast".
>
> People like that absolutely hated the traditional SCSI disk spin-up times, and they found fsck times
> completely unacceptable, and they want the RAM to be scrubbed and initialized incrementally so that you
> can boot and get the system up and the loads started ASAP. Keep scrubbing and initializing more RAM in
> parallel, because it's going to take a while before you're under full load and need it all anyway.
>
> Those people tend to be a fairly special breed.
>
> There's very much the other kind of server person, who is used to (and so resigned as to expect) the server
> reboot taking possibly tens of minutes, because that's just what they always did - spinning up one rotating
> platter of rust at a time to avoid the power surge of spinning them all up at the same time etc etc.
>
> I'm not even kidding. That whole "spin up one disk at a time" is
> not some exaggeration. It was standard practice in many places.
>
> Of course, it's largely historical, and had real roots - those disks used to be massive, and spinning
> them up really was a big deal. It's just that the practice of spinning disks up one at a time actually
> went on in many places long after those physically massive disk systems were long long gone.
>
> So you have this odd dichotomy of boot speed expectations in the big iron world. Some want it
> to be really instant and quick. And some are perfectly happy with ridiculously slow boots.
>
> I've had access to big powerful hardware that was almost entirely useless for kernel development, because
> while they could compile a new kernel really quickly, they then took literally several minutes to boot
> the damn thing. So a much slower machine was actually much better at doing any actual testing on.
>
Yeah, I had the same issue with a HP-UX machine (HP-747i) in the 90s for developing for a 68030 based VME communication card. Compiling and linking the kernel driver took seconds, rebooting without fsck took 15 minutes, with fsck 25 and it was really easy to crash the machine when communicating over the VME bus. What a luck that I could do most of the 030 part of the code on my Atari TT, otherwise I would probably not have survived that project.
Topic | Posted By | Date |
---|---|---|
CPU & Memory bit flips | Ganon | 2021/03/03 10:05 AM |
Also "Silent Data Corruption" | Adrian | 2021/03/03 11:42 AM |
Thanks for the reference | Ganon | 2021/03/03 12:47 PM |
Implications for linux page cache | anon | 2021/03/03 12:54 PM |
Implications for linux page cache | Linus Torvalds | 2021/03/03 02:54 PM |
memory errors | blaine | 2021/03/03 03:53 PM |
memory errors | anon2 | 2021/03/03 06:30 PM |
memory errors | dmcq | 2021/03/04 06:16 AM |
memory errors | Etienne Lorrain | 2021/03/04 07:26 AM |
memory errors | dmcq | 2021/03/04 07:40 AM |
memory errors | Etienne Lorrain | 2021/03/04 07:58 AM |
memory errors | dmcq | 2021/03/04 08:12 AM |
memory errors | Carson | 2021/03/05 03:31 AM |
memory errors | Etienne Lorrain | 2021/03/05 07:23 AM |
memory errors | rwessel | 2021/03/05 08:48 AM |
memory errors | dmcq | 2021/03/05 01:01 PM |
memory errors | rwessel | 2021/03/05 01:23 PM |
memory errors | dmcq | 2021/03/05 01:51 PM |
memory errors | Brendan | 2021/03/06 12:38 AM |
memory errors | Carson | 2021/03/06 02:35 AM |
memory errors | Carson | 2021/03/06 07:24 AM |
memory errors | David Hess | 2021/03/04 02:44 PM |
memory errors | rwessel | 2021/03/04 06:14 PM |
memory errors | Linus Torvalds | 2021/03/04 09:21 PM |
memory errors | anon2 | 2021/03/04 10:46 PM |
memory errors | Carson | 2021/03/05 03:43 AM |
memory errors | anon2 | 2021/03/05 08:55 AM |
memory errors | gallier2 | 2021/03/05 03:22 AM |
memory errors | dmcq | 2021/03/05 01:59 PM |
memory errors | David Hess | 2021/03/06 05:27 AM |
memory errors | Carson | 2021/03/06 07:44 AM |
memory errors | Gabriele Svelto | 2021/03/06 11:11 AM |
memory errors | David Hess | 2021/03/06 11:28 AM |
memory errors | Michael S | 2021/03/06 03:45 PM |
memory errors | Doug S | 2021/03/04 11:48 AM |
memory errors | Michael S | 2021/03/04 12:36 PM |
memory errors | Jörn Engel | 2021/03/04 04:32 PM |
memory errors | Linus Torvalds | 2021/03/04 09:47 PM |
memory errors | Etienne Lorrain | 2021/03/05 02:09 AM |
memory errors | Michael S | 2021/03/05 05:06 AM |
memory errors | Linus Torvalds | 2021/03/05 12:59 PM |
memory errors | rwessel | 2021/03/05 01:32 PM |
memory errors | rwessel | 2021/03/05 01:37 PM |
memory errors | zArchJon | 2021/03/06 09:39 PM |
memory errors | Gabriele Svelto | 2021/03/06 01:58 PM |
memory errors | Jörn Engel | 2021/03/05 11:12 AM |
Amiga recoverable RAM disk? | Carson | 2021/03/05 04:03 AM |
Thanks - TIL a cool Amiga feature (nt) (NT) | John | 2021/03/05 01:51 PM |
Another cool Amiga feature, datatypes | Charles | 2021/03/06 01:01 AM |
Another cool Amiga feature, datatypes | Jukka Larja | 2021/03/06 02:23 AM |
Another cool Amiga feature, datatypes | Anon | 2021/03/06 01:40 PM |
Another cool Amiga feature, filesystems | Marcus | 2021/03/07 01:28 AM |
CPU & Memory bit flips | zArchJon | 2021/03/04 07:39 AM |
CPU & Memory bit flips | dmcq | 2021/03/04 07:59 AM |
CPU & Memory bit flips | rwessel | 2021/03/04 01:27 PM |
speak of the devil | Robert Williams | 2021/03/05 08:53 AM |
speak of the devil | dmcq | 2021/03/05 12:26 PM |
speak of the devil | Robert Williams | 2021/03/05 04:15 PM |