FWIW,  I dipped into the higher levels of real-time-systems development several times in my career.  The earliest being a control system (circa 1981) for the LANL Proton Storage Ring where one naturally can't afford anything *but* failsafe implementations, etc. The stakes are just too 'ffing high and the coupling to electrooptomechanical systems quite intimate.

The "digital" components of such systems might have had the opportunity to ignore timing issues and simply "execute the same steps" regardless of timing.  But in fact many software-driven (sub)systems represented time-critical processes and sometimes were up agains the timing limits of the analog components which had no leeway in their "execution".

There are all kinds of analogies in federated (distributed) simulation environments which Glen (and others here) probably know much better than I, where different "clocks" matter, and different levels of synchronization and reproducibility are in play.   The Postscript interpreters, printers, and film recorders were also pseudo real-time systems since some of the timing components were in fact software controlled (for example, the film recorders were "stroke" devices with software driving D-A converters to "sweep" out vectors and "clip" the on/off of the beam with appropriate analog component delays/biases/gains needing to be calibrated for.   Fortunately failures in this step did not (usually) damage anyone or risk anyone's health and safety (like the beam in the PSR did).

Regarding identity and equivalence, I prefer the phrase: "close enough for who it's for"...


On 10/21/22 11:18 AM, glen wrote:
Ha! If we're going to argue about words, then let's stick with the word "identity" and skip the "metaphor" nonsense. You and Frank seem to be using the word in a weird way. Identity means "the exact same particular thing over any differencing available" or somesuch. I mean, it's used that way in phrases like "identity theft" as well as mathematical identity. It's equivalence sets all the way down. I just can't imagine any working computationalist would ever say anything like "executed identically" unless ... well ... the exact same process, with the exact same steps, happened.

I suppose there are deep philosophical intuitions pried at by the words "emulation" versus "simulation". And one can argue (again with help from Christian List) about whether there exist fully closed ontological walls like we try to create with things like Jails, HyperV, Docker, VM's like Java's, etc. But "execute identically" is a phrase that would only be used by someone who worked *way* above such levels (assuming levels even exist at all). It's a bit like talking to the kids programming websites these days, with access to infinite disk space, infinite memory, steeped in continuous delivery, etc. [⛧]

Layers of abstraction are fine. Use 'em when you need 'em. But we shouldn't posture by invoking things like "instruction sets" and "execute identically" in the same breath. (Not that you did that ... just sayin'.)


[⛧] Rant: This is a good talk <https://www.youtube.com/watch?v=8Ab3ArE8W3s>. But I get super irritated when people use *toy* code in their rhetoric and leave large scale deployment as an exercise for the reader. Yeah, fine. The REPL is cool and all. But when my simulation takes a fvcking WEEK to execute, it's difficult to sympathize. I've recently been playing around with VSCodium, which is pretty cool. But whatever, man. I still have to upload the code somewhere and execute it. Get off my lawn!

On 10/21/22 09:24, Steve Smith wrote:

As a counter-example,  we ran film recorders whose "guts" were built by Ed Fredkin's Information International company and were built to the spec of Dec PDP-11 (I think 11?) and it was anecdotally agreed among the user community (of a few thousand delivered units in the world?) that these PDP-clones *never* failed to execute the code identically to the machines they were patterned after.   I don't remember the details of implementation of these 70's era hardware designs, but I understood that they III designed their own PCBs but (obviously?) used the same CPU chips... I don't know about all the other support components... A likely answer to this pondering is that these machines did not run a general purpose OS and the III software/system people probably made up for any differences in Software/Timing/Error Handling?

If Owen is listening in here, I think he was there for more than a little of this from inside Apple/Sun?

- Steve

PS.   To concede/confront glen's sentiment that: " 'Metaphor' is an evil word, used only by manipulators and gaslighters",   I would offer that the use of *conceptual metaphor*  is to thinking as noise is to simulated annealing, and his point about "tighter or looser equivalence" might well be the best argument *for* the use of metaphorical thinking?  I can't believe I'm stirring/kicking this can of worm-hornets down the street again...



-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom 
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives:  5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
 1/2003 thru 6/2021  http://friam.383.s1.nabble.com/

Reply via email to