Ok. I understand.

Well this eartly salvation is not only typical of AI. Voegelin coined the
term "inmanentization of the eschaton" as the common caracteristic of the
modern political and scientific movements.


which means the belief of a heaven on earth accesible by means of some form
of hidden knowledge, that make people transcend the reality towards another
reality free of suffering and contradictions. That knowledge could be
transfered  generation by generation (masons, gnostics etc)  or
discovered/read in nature by choosen people. This knowledge can have a
theological nature  a philosophical nature (marxists, niettzcheans,
progressivists) or a scientific nature (scientists).

Voegelin say that what drives modernity in the Western world is -more or
less- the desire of the Christian eschatology and the despair of it.  the
desire for salvation and the impatience with the Christian parousía (the
end of the time). This impatience  generates heresy after heresy each one
more radical and promises a more close coming of heaven in Earth. According
with voegelin, there is a continuum between the ancient gnostics and the
modern uthopias, and share the same basic impulses.

So this apocaliptic AI is one more scientist wave in this vast historical
process.  Almost  every scientific discipline has promised a kind of
salvation for himself.

2013/3/22 Evgenii Rudnyi <use...@rudnyi.ru>

> On 22.03.2013 15:31 Alberto G. Corona said the following:
>  I a previous discussion in this list ("robotic truth")  I argue that
>> a robot in a competitive environment has to develop (or be
>> programmed for) all the elements of  religión and beliefs to
>> cooperate and survive.
> The goal of the book of Robert Geraci is different. He is after all a
> professor of religious studies. His goal were rather to compare typical
> Christian ideas of Apocalypse with the ideas that one can find in the AI
> community. The book as such is academic in nature and as a result a bit
> boring. Yet, he is convincing and the book seems justifie what he writes in
> the introduction:
> p. 7 "Apocaliptic AI is a powerful reconciliation of religion and science.
> The sacred categories of Jewish and Christian apocalyptic traditions have
> thoroughly penetrated the futuristic musings of important researches in
> robotics and artificial intelligence. Those categories have serious effects
> in robotics research, virtual reality/online gaming, and contemporary
> disputes over the nature of consciousness and personhood, public policy,
> and theology."
> The author seemingly enjoys it. The book could though be a warning for
> atheists who fight with Church and forget to check what one can find among
> AI studies.
> Evgenii
>> 2013/3/22 Evgenii Rudnyi <use...@rudnyi.ru>
>>  On 22.03.2013 13:47 Stephen P. King said the following:
>>>  On 3/22/2013 7:16 AM, Evgenii Rudnyi wrote:
>>>>  Quotes from Robert Geraci, Apocalyptic AI: Visions of Heaven
>>>>> in Robotics, Artificial Intelligence, and Virtual Reality
> --
> You received this message because you are subscribed to the Google Groups
> "Everything List" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to 
> everything-list+unsubscribe@**googlegroups.com<everything-list%2bunsubscr...@googlegroups.com>
> .
> To post to this group, send email to 
> everything-list@googlegroups.**com<everything-list@googlegroups.com>
> .
> Visit this group at 
> http://groups.google.com/**group/everything-list?hl=en<http://groups.google.com/group/everything-list?hl=en>
> .
> For more options, visit 
> https://groups.google.com/**groups/opt_out<https://groups.google.com/groups/opt_out>
> .


You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to everything-list+unsubscr...@googlegroups.com.
To post to this group, send email to everything-list@googlegroups.com.
Visit this group at http://groups.google.com/group/everything-list?hl=en.
For more options, visit https://groups.google.com/groups/opt_out.

Reply via email to