comments below...
[BG]
Hi,
Your philosophical objections aren't really objections to my perspective, so
far as I have understood so far...
[TS]
Agreed. They're to the Eliezer perspective that Vlad is arguing for.
[BG]
I don't plan to hardwire beneficialness (by which I may not mean precisely the
same thing as "Friendliness" in Eliezer's vernacular), I plan to teach it ...
to an AGI with an architecture that's well-suited to learn it, by design...
[TS]
This is essentially what we do with our kids, so no objections to the
methodology here. But from the "you have to guarantee it or we're doomed"
perspective, that's not good enough.
[BG]
I do however plan to hardwire **a powerful, super-human capability for
empathy** ... and a goal-maintenance system hardwired toward **stability of
top-level goals under self-modification**. But I agree this is different from
hardwiring specific goal content ... though it strongly *biases* the system
toward learning certain goals.
[TS]
Hardwired empathy strikes me as a basic oxymoron. Empathy must involve embodied
experience and the ability to imagine the embodied experience of another. When
we have an empathic experience, it's because we see ourselves in another's
situation - it's hard to understand what empathy could mean without that basic
subjective aspect.
-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51
Powered by Listbox: http://www.listbox.com