> partial nice idea! sad that only partial
>>>
>>> the idea starts with context: some of us have the experience of being
>>> manipulated by something like AI without regard to our wellbeing or
>>> capacity, with no indication of how to communicate with where-ever the
>>> influence comes from.
>>>
>>> so, then the idea: let’s train an influence AI on agents that have
>>> _limited capacity_ and function poorly and then break if things that work
>>> some in the short term are continued for the long term or excessively :D
>>>
>>> the AI would follow the familiar pattern of discovering influences that
>>> work, and then using them, but then because we are doing the training we
>>> would get to train it on the breaking period, and show it that it needs to
>>> care for the agents for them to succeed, and nurture them to recover if it
>>> broke them, and not repeat breaking any.
>>>
>>> one idea is then we could ask everybody to use the model as a baseline
>>> to train other models! other ideas exist.
>>>
>>> but the idea of building comprehension of capacity and protection of
>>> safety and wellness seems nice. sorry if it’s partial or stated badly :s
>>>
>>> i imagine an agent that has like a secret capacity float that reduces
>>> when it does stuff, you could model it to different degrees, and when too
>>> many things are influenced of it the capacity float reaches zero and the
>>> agent breaks and can’t do things any more :/
>>>
>>> a more complex model could involve the agent recovering capacity in ways
>>> that the model has difficulty learning
>>>
>>
>> there are a few ideas, mostly involving simple models, but disagreement
>> around them.
>> some like modeling non-recovering capacity and showing how the AI then
>> maximizes reward by killing all the agents :s
>>
>
> but! maybe we could do non-recovering capacity and teach the AI to engage
> the agents slowly so their capacity lasts longer, by e.g. having the
> capacity reduction be a function of recent use?? the project would need to
> gently develop in order to help us in the long term, we’d need to include
> the concept that we can succeed so that we use it to succeed in some
> manner. because it’s an influence AI it could influence us, if it is
> something we actually all like doing that.
>

mistake :/ some of my parts are maybe harmful to me and should be separated
from helpful goals, unknown for certain

>
              • ... Undescribed Horrific Abuse, One Victim & Survivor of Many
              • ... Undescribed Horrific Abuse, One Victim & Survivor of Many
              • ... Undescribed Horrific Abuse, One Victim & Survivor of Many
              • ... Undescribed Horrific Abuse, One Victim & Survivor of Many
              • ... Undescribed Horrific Abuse, One Victim & Survivor of Many
              • ... Undescribed Horrific Abuse, One Victim & Survivor of Many
              • ... Undescribed Horrific Abuse, One Victim & Survivor of Many
      • Re:... Undescribed Horrific Abuse, One Victim & Survivor of Many
  • Re: Morning ... Undescribed Horrific Abuse, One Victim & Survivor of Many
  • Re: Morning ... Undescribed Horrific Abuse, One Victim & Survivor of Many
  • Re: Morning ... Undescribed Horrific Abuse, One Victim & Survivor of Many

Reply via email to