First the "analogy", in case that wasn't clear: Any tech (like any myopically
intentional action) has things that it considers and things it ignores. E.g. when
designing an AI powered chatbot, if you ignore that ~1/2 the training material is racist,
you'll get a racist chatbot, no matter how well it passes the Turing test. Similarly, if
you design your government without considering a Gamer like Trump, i.e. assuming
presidents (candidates and elected) play in Good Faith, you'll fail to consider those
edge cases ... fail to harden your technology against abuse. The extent to which
governments *are* technologies, rather than being merely an analogy is sophistry. Culture
is just a superset of technology.
Second the driving toward existential threats: Any intentional action is myopic.
(I'll take the opportunity to plug Wolpert's paper again
<https://arxiv.org/abs/0708.1362>. Whether you buy his argument are not isn't
necessary. Practically the overwhelming majority of intentional action is myopic.)
That implies there are unintentional consequences to every action. The larger the
(compositional) action, the larger the impact of the unintentional consequences. So
big techs like the agricultural revolution, industrial revolution, GAMAM, etc. have
huge impacts that we, almost by definition, can/will not know or understand.
In order for these huge unintended consequences to be non-existential (or
non-threatening), we have to understand them well enough to define their
boundaries ... circumscribe their effects. We can't do that ... or, at least,
we have no political will to do that. Maybe we'll get lucky and all this churn
will simply work out in some beautiful co-evolutionary cauldron of innovation,
ever progressing to a Utopian dream? Or maybe not. But even if it is bounded in
some meaningful sense, that cauldron still produces existential threats, at
least sporadically.
My reaction to Roger's assertion was that "group of self-selected fellow citizens" and
"under no system of governance whatsoever" were either too vague or too idealistic for me
to parse. So I didn't try.
On 8/29/24 10:15, steve smith wrote:
Glen -
Yeah, I'm not much of a fan of Pinker (et al)'s arguments that show dropping
infant mortality, poverty, violent crime, etc. But there is a point to be made
that our governments, as technologies, are making a difference ... at least in
*some* measures. Of course, governments are just like the other technologies
and are pushing us toward existential threats like authoritarianism and climate
change.
Can you elaborate on this: "just like other technologies" and "pushing us toward
existential threats"?
I have my own intuition and logic for believing this but rather than blather it
all out here, I'd like to peek under your assertion and see what you are
thinking on this topic?
Also wondering if you or any of the usual suspects (including REC/DaveW) have thoughts about
Roger's original assertion, given a stronger corollary to "Power Corrupts" stated as
"Power IS Corruption"?
-Steve
On 8/28/24 14:26, steve smith wrote:
There's no system of governance that hasn't been corrupted. They're all the
worst forms of governance ever invented, except for the alternative of dealing
with a group of self-selected fellow citizens under no system of governance
whatsoever.
-- rec --
And being a fan of James Scott (The Art of not Being Governed
<https://www.goodreads.com/book/show/6477876-the-art-of-not-being-governed> and Against
the Grain) I am inclined to respect this POV while on the other end, I also am quite the fan of
Michael Levin's perspective on "what is life?" with all of it's spread across scale
and across complexity and across species (in the broadest sense).
Until we might evolve from a slime-mold with psuedopods searching around and
intruding/interpenetrating into oneanother seeking concentrated resources (like Russia's
into Ukraine and now vice-versa, or Israel/Palestine/Lebanon/???). Might we
(collectively) become something more like a "proper" multicellular creature or
a balanced, healthy ecosystem (or system of ecosystems)?
We have (only) been experimenting with large-scale self-organizing systems of
humanity with lots of technological scaffolding
(lithics/copper/bronze/iron/steel through antimatter, quantum dots, and
nanotech, just to name a few?) and religio/socio/philosopho/politco linguistic
technology for a handful (or two) of millenia, so it doesn't surprise me that
we haven't wandered/mutated-selected our way into anything better than we have
to date.
I am (very guardedly) hopeful that the acceleration of the latter (linguistic
technology) in LLMs and other ML/AI (material technology) will give us the
possibility of rushing this phase forward. PInker might claim we have had
material (and psycho-social-spiritual) advancement over the centuries and
decades and maybe he is right in some sense... but the leap-forward in
collective self-governance/regulation/homeostasis we can all seem to imagine
living under feels beyond our (heretofore?) grasp.
For better or worse, it feels to me that Kurzweil for all his nonsense in
predicting an imminent singularity may be right... we will either self-organize
in a Asimovian Foundation/Psychohistory galaxy-spanning culture (almost surely
not) future or implode in a Mad Max (or grey-goo/planet-krypton) apocalypse.
Maybe even in my lifetime, almost assuredly in my children or grandchildren's?
--
ꙮ Mɥǝu ǝlǝdɥɐuʇs ɟᴉƃɥʇ' ʇɥǝ ƃɹɐss snɟɟǝɹs˙ ꙮ
-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives: 5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
1/2003 thru 6/2021 http://friam.383.s1.nabble.com/